Posts

Resolving von Neumann-Morgenstern Inconsistent Preferences 2024-10-22T11:45:20.915Z
0.202 Bits of Evidence In Favor of Futarchy 2024-09-29T21:57:59.896Z
Pomodoro Method Randomized Self Experiment 2024-09-29T21:55:04.740Z
How Often Does Taking Away Options Help? 2024-09-21T21:52:40.822Z
Michael Dickens' Caffeine Tolerance Research 2024-09-04T15:41:53.343Z
Thoughts to niplav on lie-detection, truthfwl mechanisms, and wealth-inequality 2024-07-11T18:55:46.687Z
An AI Race With China Can Be Better Than Not Racing 2024-07-02T17:57:36.976Z
Fat Tails Discourage Compromise 2024-06-17T09:39:16.489Z
Transfer Learning in Humans 2024-04-21T20:49:42.595Z
How Often Does ¬Correlation ⇏ ¬Causation? 2024-04-02T17:58:41.695Z
Collection of Scientific and Other Classifications 2024-02-15T12:58:50.626Z
Conversation Visualizer 2023-12-31T01:18:01.424Z
Please Bet On My Quantified Self Decision Markets 2023-12-01T20:07:38.284Z
Self-Blinded L-Theanine RCT 2023-10-31T15:24:57.717Z
Examples of Low Status Fun 2023-10-10T23:19:26.212Z
Precision of Sets of Forecasts 2023-09-19T18:19:18.053Z
Have Attention Spans Been Declining? 2023-09-08T14:11:55.224Z
The Evidence for Question Decomposition is Weak 2023-08-28T15:46:31.529Z
If I Was An Eccentric Trillionaire 2023-08-09T07:56:46.259Z
Self-Blinded Caffeine RCT 2023-06-27T12:38:55.354Z
Properties of Good Textbooks 2023-05-07T08:38:05.243Z
Iqisa: A Library For Handling Forecasting Datasets 2023-04-14T15:16:55.726Z
Subscripts for Probabilities 2023-04-13T18:32:17.267Z
High Status Eschews Quantification of Performance 2023-03-19T22:14:16.523Z
Open & Welcome Thread — March 2023 2023-03-01T09:30:09.639Z
Open & Welcome Thread - December 2022 2022-12-04T15:06:00.579Z
Open & Welcome Thread - Oct 2022 2022-10-02T11:04:06.762Z
Turning Some Inconsistent Preferences into Consistent Ones 2022-07-18T18:40:02.243Z
What is Going On With CFAR? 2022-05-28T15:21:51.397Z
Range and Forecasting Accuracy 2022-05-27T18:47:44.315Z
scipy.optimize.curve_fit Is Awesome 2022-05-07T10:57:28.278Z
Brain-Computer Interfaces and AI Alignment 2021-08-28T19:48:52.614Z
Open & Welcome Thread – April 2021 2021-04-04T19:25:09.049Z
An Exploratory Toy AI Takeoff Model 2021-01-13T18:13:14.237Z
Cryonics Cost-Benefit Analysis 2020-08-03T17:30:42.307Z
"Do Nothing" utility function, 3½ years later? 2020-07-20T11:09:36.946Z
shortplav 2020-06-20T21:15:06.105Z

Comments

Comment by niplav on Stupid Question: Why am I getting consistently downvoted? · 2024-12-08T23:16:17.813Z · LW · GW

Has @MadHatter replied or transferred the money yet?

Comment by niplav on O O's Shortform · 2024-12-08T23:12:17.702Z · LW · GW

Not OP but it could be that o1 underperformed their expectation.

Comment by niplav on Hazard's Shortform Feed · 2024-12-08T22:48:10.294Z · LW · GW

Thank you, that's useful evidence!

Comment by niplav on Hazard's Shortform Feed · 2024-12-08T00:38:30.579Z · LW · GW

Thank you for collecting those links :-)

I've listened to two or three of the interviews (and ~three other talks from a long time ago), and I still have no clue what the central claims are, what the reasoning supporting them is &c. (I understand it most for Zvi Mowshowitz and Sarah Constantin, less for Jessica Taylor, and least for Benjamin Hoffman & Vassar). I also don't know of anyone who became convinced of or even understood any of Michael Vassar's views/stances through his writing/podcasts alone—it appears to almost always happens through in-person interaction.

Comment by niplav on Algebraic Linguistics · 2024-12-08T00:04:40.186Z · LW · GW

See also Latin, Greek and other letters used in math, science and engineering on Wikipedia.

Comment by niplav on Open Thread Fall 2024 · 2024-12-05T17:12:16.313Z · LW · GW

The obvious advice is of course "whatever thing you want to learn, let an LLM help you learn it". Throw that post in the context window, zoom in on terms, ask it to provide examples in the way the author intended it, let it generate exercises, let it rewrite it for your reading level.

If you're already doing that and it's not helping, maybe… more dakka? And you're going to have to expand on what your goals are and what you want to learn/make.

Comment by niplav on (The) Lightcone is nothing without its people: LW + Lighthaven's big fundraiser · 2024-11-30T21:38:08.497Z · LW · GW

I remember that Lightcone was interested in working on human intelligence amplification and/or pausing AI (I can't find the LW comment, I'm afraid). Is that still part of the plan?

Comment by niplav on OpenAI Email Archives (from Musk v. Altman and OpenAI blog) · 2024-11-28T23:05:20.872Z · LW · GW

Huh, thanks, nevermind.

Comment by niplav on OpenAI Email Archives (from Musk v. Altman and OpenAI blog) · 2024-11-28T22:27:12.625Z · LW · GW

We had originally just wanted space cycles donated

I think this is a mistake, and it should be "spare cycles" instead.

Comment by niplav on Causal inference for the home gardener · 2024-11-27T23:26:38.047Z · LW · GW

I think I'd've wanted to know about tigramite when learning about causal inference, it's a library for doing causal inference on time-series data.

Comment by niplav on Epistemic status: poetry (and other poems) · 2024-11-21T21:00:12.231Z · LW · GW

Epistemic status: I saw the best minds of my generation…
Epistemic status: Proven, by me

Comment by niplav on What are the good rationality films? · 2024-11-21T11:38:21.228Z · LW · GW

Datapoint: I didn't enjoy margin call, because it didn't try to explain the crisis, and the character of the CEO was deliberately dumbed down in a way that I don't think real finance CEOs are.

Comment by niplav on Announcing turntrout.com, my new digital home · 2024-11-17T18:16:32.823Z · LW · GW

I remember writing a note a few years ago on who I wished would to create a long site, and your pseudonym was on the list. Happy to see that this has happened, even if for unfortunate reasons.

Comment by niplav on shortplav · 2024-11-15T20:45:15.957Z · LW · GW

There are two pre-existing Manifold Markets questions on whether LLM scaling laws will hold until 2027 and 2028, respectively, with currently little trading volume.

Comment by niplav on Resolving von Neumann-Morgenstern Inconsistent Preferences · 2024-11-13T09:16:37.077Z · LW · GW

Not published anywhere except here and on my site.

Comment by niplav on Alex K. Chen's Shortform · 2024-11-11T08:15:47.052Z · LW · GW

I guess a problem here would be the legal issues around checking whether the file is copyright-protected.

Comment by niplav on AI #89: Trump Card · 2024-11-08T14:08:07.759Z · LW · GW

Finally, note to self, probably still don’t use SQLite if you have a good alternative? Twice is suspicious, although they did fix the bug same day and it wasn’t ever released.

SQLite is well-known for its incredibly thorough test suite and relatively few CVEs, and with ~156kloc (excluding tests) it's not a very large project, so I think this would be an over-reaction. I'd guess that other databases have more and worse security vulnerabilities due to their attack surface—see MySQL with its ~4.4mloc (including tests). Big Sleep was probably now used on SQLite because it's a fairly small project of which large parts can fit into an LLMs' context window.

Maybe someone will try to translate the SQLite code to Rust or Zig using LLMs—until then we're stuck.

Comment by niplav on avturchin's Shortform · 2024-10-31T07:59:02.491Z · LW · GW

Not surprising to me: I've lived in a city with many stray dogs for less than half a year, and got "attacked" ("harrassed" is maybe a better term) by a stray dog twice.

Comment by niplav on (i no longer endorse this post) - cryonics is a pascal's mugging? · 2024-10-26T08:45:43.340Z · LW · GW

Not a Pascal's mugging to the best of my knowledge.

Comment by niplav on Sodium's Shortform · 2024-10-26T00:50:06.054Z · LW · GW

I think normally "agile" would fulfill the same function (per its etymology), but it's very entangled with agile software engineering.

Comment by niplav on Zach Stein-Perlman's Shortform · 2024-10-25T16:00:06.930Z · LW · GW

Two others that come to mind:

  • Metaculus (used to be better though)
  • lobste.rs (quite specialized)
  • Quanta Magazine has some good comments, e.g. this article has the original researcher showing up & clarifying some questions in the comments
Comment by niplav on shortplav · 2024-10-24T08:56:53.966Z · LW · GW

Apparently a Thompson-hack-like bug occurred in LLVM (haven't read the post in detail yet). Interesting.

Comment by niplav on Resolving von Neumann-Morgenstern Inconsistent Preferences · 2024-10-22T11:50:08.306Z · LW · GW

Submission statement: I mostly finished this a year ago, but held off on posting because I was planning on improving it and writing a corresponding "here's the concepts without the math" post. Might still happen, but now I'm not aiming at a specific timeline.

Things I now want to change:

  • Soften the confidence in the vNM axioms, since there's been some good criticisms
  • Revamp the whole ontological crisis section to be more general
  • Rewrite from academese to easier
  • Move proofs to an appendix
  • Create some manim videos to illustrate
  • Merge with this post
  • Many other things

Still, I hope this is kinda useful for some people.

Edit: Also, there's some issues with the MathJax and dollar signs, I will fix this later.

Comment by niplav on yams's Shortform · 2024-10-20T07:18:59.559Z · LW · GW

Apologies for the soldier mindset react, I pattern-matched to some more hostile comment. Communication is hard.

Comment by niplav on yams's Shortform · 2024-10-19T21:11:08.974Z · LW · GW

Grants to Redwood Research, SERI MATS, NYU alignment group under Sam Bowman for scalable supervision, Palisade research, and many dozens more, most of which seem net positive wrt TAI risk.

Comment by niplav on Most arguments for AI Doom are either bad or weak · 2024-10-14T17:53:37.226Z · LW · GW

Yudkowsky 2017, AronT 2023 and Gwern 2019, if you're curious why you're getting downvoted.

(I tried to figure out whether this method of estimation works, and it seemed more accurate than I thought, but then I got distracted).

Comment by niplav on An AI Race With China Can Be Better Than Not Racing · 2024-10-14T17:46:23.401Z · LW · GW

There's two arguments you've made, one is very gnarly, the other is wrong :-):

  1. "the sheer number of parameters you have chosen arbitrarily or said "eh, let's assume this is normally distributed" demonstrates the futility of approaching this question numerically."
  2. "simply stating your preference ordering"

I didn't just state a preference ordering over futures, I also ass-numbered their probabilities and ass-guessed ways of getting there. For to estimate an expected value of an action, one requires two things: A list of probabilities, and a list of utilities—you merely propose giving one of those.

(As for the "false precision", I feel like the debate has run its course; I consider Scott Alexander, 2017 to be the best rejoinder here. The world is likely not structured in a way that makes trying harder to estimate be less accurate in expectation (which I'd dub the Taoist assumption, thinking & estimating more should narrow the credences over time. Same reason why I've defended the bioanchors report against accusations of uselessness with having distributions over 14 orders of magnitude).

Comment by niplav on Dalcy's Shortform · 2024-10-05T11:37:34.928Z · LW · GW

The way I do this is use the Print as PDF functionality in the browser on every single post, and then concatenate them using pdfunite.

Comment by niplav on AI #84: Better Than a Podcast · 2024-10-04T10:01:56.791Z · LW · GW
  • Building a superintelligence under current conditions will turn out fine.
  • No one will build a superintelligence under anything like current conditions.
  • We must prevent at almost all costs anyone building superintelligence soon.

I don't think this is a valid trilemma: Between fine and worth preventing at "almost all costs" there is a pretty large gap. I think "fine" was intended to mean "we don't all die" or something as bad as that.

Comment by niplav on Pomodoro Method Randomized Self Experiment · 2024-09-30T09:02:11.704Z · LW · GW

Thanks, I'll improve the data and then analyse it when I have more time.

Comment by niplav on Nathan Young's Shortform · 2024-09-24T11:33:59.882Z · LW · GW

Relevant: When pooling forecasts, use the geometric mean of odds.

Comment by niplav on How Often Does Taking Away Options Help? · 2024-09-23T12:18:50.814Z · LW · GW

You're right. I'll rerun the analysis and include 2x2 games as well.

Comment by niplav on shortplav · 2024-09-23T09:16:14.975Z · LW · GW

This is interesting, thank you—I hadn't considered the case where an existing contract needs to be renewed.

I wonder why under your understanding predicts stagnating or decreasing salaries in this world? Currently, employees sometimes quit if they haven't gotten a raise in a while, and go to other companies where they can earn more. In this mechanism, this can be encoded as choosing a higher , which is set just at the level where the employee would be indifferent between staying at the company and going to job-hunt again.

I agree that this would have downsides for candidates with few other options, and feel a bit bad about that. Not sure whether it's economically efficient, though.

Comment by niplav on shortplav · 2024-09-23T09:09:24.123Z · LW · GW

The question is, would it be better for companies than the current situation? Because it's the company who decides the form of the interview, so if the answer is negative, this is not going to happen.

Yeah, I don't think this is going to be adopted very soon. My best guess at how that could happen is if people try it in low-stakes contexts where the parties are ~symmetric in power, and this then spreads through e.g. people who do consulting for small startups, to salaries for high-value employees in small startups, to salaries for high-value employees in general etc.

Another way this could happen is if unions push for it, but I don't see that happening anytime soon.

(I'm going to see whether me putting this up as a way of determining rates can work, but probably not.)

Comment by niplav on shortplav · 2024-09-23T09:04:31.160Z · LW · GW

Yeah, the spherical cow system would be using the VCG mechanism with the Clarke pivot rule, but that would usually require some subsidy. There can be no spherical cow system which elicits truthful bids without subsidy, sadly :-/.

Comment by niplav on shortplav · 2024-09-23T09:02:05.623Z · LW · GW

Does this require some sort of enforcement mechanism to ensure that neither party puts in a bad-faith bid as a discovery mechanism for what number to seek in their real negotiations?

Maybe there's a misunderstanding here—the mechanism I was writing about would be the "real negotiations" (whatever result falls out of the mechanism now is what's going to happen). As in, there can be a lot of talking about salaries before the this two-sided sealed auction is performed, but the salary is decided through the auction.

I know of some software engineers who have published their salary history online.

Comment by niplav on How Often Does Taking Away Options Help? · 2024-09-23T08:33:52.786Z · LW · GW

Maybe because the URL is an http URL instead of https.

Comment by niplav on Sodium's Shortform · 2024-09-21T11:36:00.222Z · LW · GW

When will this be revealed?

Comment by niplav on Bogdan Ionut Cirstea's Shortform · 2024-09-18T17:51:13.578Z · LW · GW

I also have this market for GPQA, on a longer time-horizon:

https://manifold.markets/NiplavYushtun/will-the-gap-between-openweights-an

Comment by niplav on Hyperpolation · 2024-09-16T18:45:13.403Z · LW · GW

I'm surprised that the paper doesn't mention analytic continuations of complex functions—maybe that is also taken as an instance of extrapolation?

Comment by niplav on shortplav · 2024-09-08T19:30:25.551Z · LW · GW

The current state of the art for salary negotiations is really bad. It rewards disagreeableness, stubornness and social skills, and is just so unelegant.

Here's a better way of doing salary negotiation:

Procedure via a two-sided sealed-bid auction, splitting the difference in bids[1]:

  • Normal interviewing happens.
  • Job-seeker decides on their minimum acceptable rate .
  • Employee-seeker decides on the maximum acceptable payment .
  • Both reveal and , either first through hashsums of the numbers (with random text appended) and then the cleartext, or simply at the same time.
    • If , then the rate is (i.e. the mean of the two bids).
    • If , then no agreement takes place and both parties fall back to their BATNAs.

and do not need to be positive! It might be that the potential employee likes the project so much that they set to zero or even negative—an exceptionally great idea might be worth paying for. Or might be negative, in that case one party would be selling something.

I'm not aware of anyone proposing this kind of auction for salary negotiation in particular, Claude 3.5 Sonnet states that it's similar to Vickrey auctions, but in this case there is no second price, and both parties are symmetrical.


  1. I think that the setup described is probably not incentive-compatible due to the Myerson-Satterthwaite theorem, like the first-price sealed-bid auction. (I still think it's a vast improvement over the current state of the art, however). For an incentive-compatible truthful mechanism the Vickrey-Clark-Groves mechanism can be used, but I'm still a bit unsure how the subsidising would work. ↩︎

Comment by niplav on RussellThor's Shortform · 2024-09-06T04:23:18.188Z · LW · GW

Thanks, that updates me. I've been enjoying your well-informed comments on big training runs, thank you!

Comment by niplav on RussellThor's Shortform · 2024-09-05T20:50:01.903Z · LW · GW

On priors I think that Google Deepmind is currently running the biggest training run.

Comment by niplav on Thoughts to niplav on lie-detection, truthfwl mechanisms, and wealth-inequality · 2024-09-05T01:50:59.139Z · LW · GW
Comment by niplav on shortplav · 2024-09-04T21:43:51.107Z · LW · GW

The Variety-Uninterested Can Buy Schelling-Products

Having many different products in the same category, such as many different kinds of clothes or cars or houses, is probably very expensive.

Some of us might not care enough about variety of products in a certain category to pay the extra cost of variety, and may even resent the variety-interested for imposing that cost.

But the variety-uninterested can try to recover some of the gains from eschewing variety by all buying the same product in some category. Often, this will mean buying the cheapest acceptable product from some category, or the product with the least amount of ornamentation or special features.

E.g. one can buy only black t-shirts and featuresless cheap black socks, and simple metal cutlery. I will, next time I'll buy a laptop or a smartphone, think about what the Schelling-laptop is. I suspect it's not a ThinkPad.

"Then let them all have the same kind of cake."

Comment by niplav on shortplav · 2024-09-04T10:09:32.723Z · LW · GW

And: yes, the games weren't normalized to be zero-sum.

Comment by niplav on shortplav · 2024-09-04T10:01:19.885Z · LW · GW

I wrote a short reply to Dagon, maybe that helps.

Otherwise I might write up a full post explaning this with examples &c.

Comment by niplav on shortplav · 2024-09-04T09:59:20.526Z · LW · GW

Updated the link to the actual code. I computed the equilibria for the full game, and then computed the payoff per equilibrium for each player, and then took the mean for each player. I did the same but with the game with one option removed. The number in the chart is the proportion of games where removing one option from player A improved the payoff (averaged over equilibria).

If the number is >0.5, then that means that for that player, removing one option from A on average improves their payoffs. (The number of options is pre-removal). I also found this interesting, but the charts are maybe a bit misleading because often removing one option from A doesn't change the equilibria. I'll maybe generate some charts for this.

I'll perhaps also write a clearer explanation of what is happening and repost as a top-level post.

Comment by niplav on shortplav · 2024-09-03T14:11:47.666Z · LW · GW

How Often Does Taking Away Options Help?

In some game-theoretic setups, taking options away from a player improves their situation. I ran a Monte-Carlo simulation to figure out how often that is the case, generating random normal form games with payoffs in , removing a random option from the first player, and comparing the Nash equilibria found via vertex enumeration of the best response polytope (using nashpy)—the Lemke-Howson algorithm was giving me duplicate results.

Code here, largely written by Claude 3.5 Sonnet.

Comment by niplav on shortplav · 2024-08-27T20:37:24.033Z · LW · GW

I find the Thompson hack very fascinating from an agent foundations perspective. It's basically a small version of reflective stability in the context of operating systems.

I used to find compilers written in their own language kind of—…distasteful, in some way? Some of that is still present, because in reality it's just that the bootstrapping chains become very long and difficult to follow. But I think a small part of that distaste was the worry that Thompson hack-style errors occur accidentally at some point, and are just propagated through the bootstrapping chain. After thinking about this for a few seconds this was of course patently ridiculous.

But under this lens reflective stability becomes really difficult, because every replicating/successor-generating subsystem needs to be adapted to have the property of reflective stability.

E.g. corrigibility is really hard if one imagines it as a type of Thompson hack, especially under relative robustness to scale. You don't just get a basin of Thompson-hackness when writing compilers and making mistakes.