Posts

Resolving von Neumann-Morgenstern Inconsistent Preferences 2024-10-22T11:45:20.915Z
0.202 Bits of Evidence In Favor of Futarchy 2024-09-29T21:57:59.896Z
Pomodoro Method Randomized Self Experiment 2024-09-29T21:55:04.740Z
How Often Does Taking Away Options Help? 2024-09-21T21:52:40.822Z
Michael Dickens' Caffeine Tolerance Research 2024-09-04T15:41:53.343Z
Thoughts to niplav on lie-detection, truthfwl mechanisms, and wealth-inequality 2024-07-11T18:55:46.687Z
An AI Race With China Can Be Better Than Not Racing 2024-07-02T17:57:36.976Z
Fat Tails Discourage Compromise 2024-06-17T09:39:16.489Z
Transfer Learning in Humans 2024-04-21T20:49:42.595Z
How Often Does ¬Correlation ⇏ ¬Causation? 2024-04-02T17:58:41.695Z
Collection of Scientific and Other Classifications 2024-02-15T12:58:50.626Z
Conversation Visualizer 2023-12-31T01:18:01.424Z
Please Bet On My Quantified Self Decision Markets 2023-12-01T20:07:38.284Z
Self-Blinded L-Theanine RCT 2023-10-31T15:24:57.717Z
Examples of Low Status Fun 2023-10-10T23:19:26.212Z
Precision of Sets of Forecasts 2023-09-19T18:19:18.053Z
Have Attention Spans Been Declining? 2023-09-08T14:11:55.224Z
The Evidence for Question Decomposition is Weak 2023-08-28T15:46:31.529Z
If I Was An Eccentric Trillionaire 2023-08-09T07:56:46.259Z
Self-Blinded Caffeine RCT 2023-06-27T12:38:55.354Z
Properties of Good Textbooks 2023-05-07T08:38:05.243Z
Iqisa: A Library For Handling Forecasting Datasets 2023-04-14T15:16:55.726Z
Subscripts for Probabilities 2023-04-13T18:32:17.267Z
High Status Eschews Quantification of Performance 2023-03-19T22:14:16.523Z
Open & Welcome Thread — March 2023 2023-03-01T09:30:09.639Z
Open & Welcome Thread - December 2022 2022-12-04T15:06:00.579Z
Open & Welcome Thread - Oct 2022 2022-10-02T11:04:06.762Z
Turning Some Inconsistent Preferences into Consistent Ones 2022-07-18T18:40:02.243Z
What is Going On With CFAR? 2022-05-28T15:21:51.397Z
Range and Forecasting Accuracy 2022-05-27T18:47:44.315Z
scipy.optimize.curve_fit Is Awesome 2022-05-07T10:57:28.278Z
Brain-Computer Interfaces and AI Alignment 2021-08-28T19:48:52.614Z
Open & Welcome Thread – April 2021 2021-04-04T19:25:09.049Z
An Exploratory Toy AI Takeoff Model 2021-01-13T18:13:14.237Z
Cryonics Cost-Benefit Analysis 2020-08-03T17:30:42.307Z
"Do Nothing" utility function, 3½ years later? 2020-07-20T11:09:36.946Z
shortplav 2020-06-20T21:15:06.105Z

Comments

Comment by niplav on Announcing turntrout.com, my new digital home · 2024-11-17T18:16:32.823Z · LW · GW

I remember writing a note a few years ago on who I wished would to create a long site, and your pseudonym was on the list. Happy to see that this has happened, even if for unfortunate reasons.

Comment by niplav on shortplav · 2024-11-15T20:45:15.957Z · LW · GW

There are two pre-existing Manifold Markets questions on whether LLM scaling laws will hold until 2027 and 2028, respectively, with currently little trading volume.

Comment by niplav on Resolving von Neumann-Morgenstern Inconsistent Preferences · 2024-11-13T09:16:37.077Z · LW · GW

Not published anywhere except here and on my site.

Comment by niplav on Alex K. Chen's Shortform · 2024-11-11T08:15:47.052Z · LW · GW

I guess a problem here would be the legal issues around checking whether the file is copyright-protected.

Comment by niplav on AI #89: Trump Card · 2024-11-08T14:08:07.759Z · LW · GW

Finally, note to self, probably still don’t use SQLite if you have a good alternative? Twice is suspicious, although they did fix the bug same day and it wasn’t ever released.

SQLite is well-known for its incredibly thorough test suite and relatively few CVEs, and with ~156kloc (excluding tests) it's not a very large project, so I think this would be an over-reaction. I'd guess that other databases have more and worse security vulnerabilities due to their attack surface—see MySQL with its ~4.4mloc (including tests). Big Sleep was probably now used on SQLite because it's a fairly small project of which large parts can fit into an LLMs' context window.

Maybe someone will try to translate the SQLite code to Rust or Zig using LLMs—until then we're stuck.

Comment by niplav on avturchin's Shortform · 2024-10-31T07:59:02.491Z · LW · GW

Not surprising to me: I've lived in a city with many stray dogs for less than half a year, and got "attacked" ("harrassed" is maybe a better term) by a stray dog twice.

Comment by niplav on (i no longer endorse this post) - cryonics is a pascal's mugging? · 2024-10-26T08:45:43.340Z · LW · GW

Not a Pascal's mugging to the best of my knowledge.

Comment by niplav on Sodium's Shortform · 2024-10-26T00:50:06.054Z · LW · GW

I think normally "agile" would fulfill the same function (per its etymology), but it's very entangled with agile software engineering.

Comment by niplav on Zach Stein-Perlman's Shortform · 2024-10-25T16:00:06.930Z · LW · GW

Two others that come to mind:

  • Metaculus (used to be better though)
  • lobste.rs (quite specialized)
  • Quanta Magazine has some good comments, e.g. this article has the original researcher showing up & clarifying some questions in the comments
Comment by niplav on shortplav · 2024-10-24T08:56:53.966Z · LW · GW

Apparently a Thompson-hack-like bug occurred in LLVM (haven't read the post in detail yet). Interesting.

Comment by niplav on Resolving von Neumann-Morgenstern Inconsistent Preferences · 2024-10-22T11:50:08.306Z · LW · GW

Submission statement: I mostly finished this a year ago, but held off on posting because I was planning on improving it and writing a corresponding "here's the concepts without the math" post. Might still happen, but now I'm not aiming at a specific timeline.

Things I now want to change:

  • Soften the confidence in the vNM axioms, since there's been some good criticisms
  • Revamp the whole ontological crisis section to be more general
  • Rewrite from academese to easier
  • Move proofs to an appendix
  • Create some manim videos to illustrate
  • Merge with this post
  • Many other things

Still, I hope this is kinda useful for some people.

Edit: Also, there's some issues with the MathJax and dollar signs, I will fix this later.

Comment by niplav on yams's Shortform · 2024-10-20T07:18:59.559Z · LW · GW

Apologies for the soldier mindset react, I pattern-matched to some more hostile comment. Communication is hard.

Comment by niplav on yams's Shortform · 2024-10-19T21:11:08.974Z · LW · GW

Grants to Redwood Research, SERI MATS, NYU alignment group under Sam Bowman for scalable supervision, Palisade research, and many dozens more, most of which seem net positive wrt TAI risk.

Comment by niplav on Most arguments for AI Doom are either bad or weak · 2024-10-14T17:53:37.226Z · LW · GW

Yudkowsky 2017, AronT 2023 and Gwern 2019, if you're curious why you're getting downvoted.

(I tried to figure out whether this method of estimation works, and it seemed more accurate than I thought, but then I got distracted).

Comment by niplav on An AI Race With China Can Be Better Than Not Racing · 2024-10-14T17:46:23.401Z · LW · GW

There's two arguments you've made, one is very gnarly, the other is wrong :-):

  1. "the sheer number of parameters you have chosen arbitrarily or said "eh, let's assume this is normally distributed" demonstrates the futility of approaching this question numerically."
  2. "simply stating your preference ordering"

I didn't just state a preference ordering over futures, I also ass-numbered their probabilities and ass-guessed ways of getting there. For to estimate an expected value of an action, one requires two things: A list of probabilities, and a list of utilities—you merely propose giving one of those.

(As for the "false precision", I feel like the debate has run its course; I consider Scott Alexander, 2017 to be the best rejoinder here. The world is likely not structured in a way that makes trying harder to estimate be less accurate in expectation (which I'd dub the Taoist assumption, thinking & estimating more should narrow the credences over time. Same reason why I've defended the bioanchors report against accusations of uselessness with having distributions over 14 orders of magnitude).

Comment by niplav on Dalcy's Shortform · 2024-10-05T11:37:34.928Z · LW · GW

The way I do this is use the Print as PDF functionality in the browser on every single post, and then concatenate them using pdfunite.

Comment by niplav on AI #84: Better Than a Podcast · 2024-10-04T10:01:56.791Z · LW · GW
  • Building a superintelligence under current conditions will turn out fine.
  • No one will build a superintelligence under anything like current conditions.
  • We must prevent at almost all costs anyone building superintelligence soon.

I don't think this is a valid trilemma: Between fine and worth preventing at "almost all costs" there is a pretty large gap. I think "fine" was intended to mean "we don't all die" or something as bad as that.

Comment by niplav on Pomodoro Method Randomized Self Experiment · 2024-09-30T09:02:11.704Z · LW · GW

Thanks, I'll improve the data and then analyse it when I have more time.

Comment by niplav on Nathan Young's Shortform · 2024-09-24T11:33:59.882Z · LW · GW

Relevant: When pooling forecasts, use the geometric mean of odds.

Comment by niplav on How Often Does Taking Away Options Help? · 2024-09-23T12:18:50.814Z · LW · GW

You're right. I'll rerun the analysis and include 2x2 games as well.

Comment by niplav on shortplav · 2024-09-23T09:16:14.975Z · LW · GW

This is interesting, thank you—I hadn't considered the case where an existing contract needs to be renewed.

I wonder why under your understanding predicts stagnating or decreasing salaries in this world? Currently, employees sometimes quit if they haven't gotten a raise in a while, and go to other companies where they can earn more. In this mechanism, this can be encoded as choosing a higher , which is set just at the level where the employee would be indifferent between staying at the company and going to job-hunt again.

I agree that this would have downsides for candidates with few other options, and feel a bit bad about that. Not sure whether it's economically efficient, though.

Comment by niplav on shortplav · 2024-09-23T09:09:24.123Z · LW · GW

The question is, would it be better for companies than the current situation? Because it's the company who decides the form of the interview, so if the answer is negative, this is not going to happen.

Yeah, I don't think this is going to be adopted very soon. My best guess at how that could happen is if people try it in low-stakes contexts where the parties are ~symmetric in power, and this then spreads through e.g. people who do consulting for small startups, to salaries for high-value employees in small startups, to salaries for high-value employees in general etc.

Another way this could happen is if unions push for it, but I don't see that happening anytime soon.

(I'm going to see whether me putting this up as a way of determining rates can work, but probably not.)

Comment by niplav on shortplav · 2024-09-23T09:04:31.160Z · LW · GW

Yeah, the spherical cow system would be using the VCG mechanism with the Clarke pivot rule, but that would usually require some subsidy. There can be no spherical cow system which elicits truthful bids without subsidy, sadly :-/.

Comment by niplav on shortplav · 2024-09-23T09:02:05.623Z · LW · GW

Does this require some sort of enforcement mechanism to ensure that neither party puts in a bad-faith bid as a discovery mechanism for what number to seek in their real negotiations?

Maybe there's a misunderstanding here—the mechanism I was writing about would be the "real negotiations" (whatever result falls out of the mechanism now is what's going to happen). As in, there can be a lot of talking about salaries before the this two-sided sealed auction is performed, but the salary is decided through the auction.

I know of some software engineers who have published their salary history online.

Comment by niplav on How Often Does Taking Away Options Help? · 2024-09-23T08:33:52.786Z · LW · GW

Maybe because the URL is an http URL instead of https.

Comment by niplav on Sodium's Shortform · 2024-09-21T11:36:00.222Z · LW · GW

When will this be revealed?

Comment by niplav on Bogdan Ionut Cirstea's Shortform · 2024-09-18T17:51:13.578Z · LW · GW

I also have this market for GPQA, on a longer time-horizon:

https://manifold.markets/NiplavYushtun/will-the-gap-between-openweights-an

Comment by niplav on Hyperpolation · 2024-09-16T18:45:13.403Z · LW · GW

I'm surprised that the paper doesn't mention analytic continuations of complex functions—maybe that is also taken as an instance of extrapolation?

Comment by niplav on shortplav · 2024-09-08T19:30:25.551Z · LW · GW

The current state of the art for salary negotiations is really bad. It rewards disagreeableness, stubornness and social skills, and is just so unelegant.

Here's a better way of doing salary negotiation:

Procedure via a two-sided sealed-bid auction, splitting the difference in bids[1]:

  • Normal interviewing happens.
  • Job-seeker decides on their minimum acceptable rate .
  • Employee-seeker decides on the maximum acceptable payment .
  • Both reveal and , either first through hashsums of the numbers (with random text appended) and then the cleartext, or simply at the same time.
    • If , then the rate is (i.e. the mean of the two bids).
    • If , then no agreement takes place and both parties fall back to their BATNAs.

and do not need to be positive! It might be that the potential employee likes the project so much that they set to zero or even negative—an exceptionally great idea might be worth paying for. Or might be negative, in that case one party would be selling something.

I'm not aware of anyone proposing this kind of auction for salary negotiation in particular, Claude 3.5 Sonnet states that it's similar to Vickrey auctions, but in this case there is no second price, and both parties are symmetrical.


  1. I think that the setup described is probably not incentive-compatible due to the Myerson-Satterthwaite theorem, like the first-price sealed-bid auction. (I still think it's a vast improvement over the current state of the art, however). For an incentive-compatible truthful mechanism the Vickrey-Clark-Groves mechanism can be used, but I'm still a bit unsure how the subsidising would work. ↩︎

Comment by niplav on RussellThor's Shortform · 2024-09-06T04:23:18.188Z · LW · GW

Thanks, that updates me. I've been enjoying your well-informed comments on big training runs, thank you!

Comment by niplav on RussellThor's Shortform · 2024-09-05T20:50:01.903Z · LW · GW

On priors I think that Google Deepmind is currently running the biggest training run.

Comment by niplav on Thoughts to niplav on lie-detection, truthfwl mechanisms, and wealth-inequality · 2024-09-05T01:50:59.139Z · LW · GW
Comment by niplav on shortplav · 2024-09-04T21:43:51.107Z · LW · GW

The Variety-Uninterested Can Buy Schelling-Products

Having many different products in the same category, such as many different kinds of clothes or cars or houses, is probably very expensive.

Some of us might not care enough about variety of products in a certain category to pay the extra cost of variety, and may even resent the variety-interested for imposing that cost.

But the variety-uninterested can try to recover some of the gains from eschewing variety by all buying the same product in some category. Often, this will mean buying the cheapest acceptable product from some category, or the product with the least amount of ornamentation or special features.

E.g. one can buy only black t-shirts and featuresless cheap black socks, and simple metal cutlery. I will, next time I'll buy a laptop or a smartphone, think about what the Schelling-laptop is. I suspect it's not a ThinkPad.

"Then let them all have the same kind of cake."

Comment by niplav on shortplav · 2024-09-04T10:09:32.723Z · LW · GW

And: yes, the games weren't normalized to be zero-sum.

Comment by niplav on shortplav · 2024-09-04T10:01:19.885Z · LW · GW

I wrote a short reply to Dagon, maybe that helps.

Otherwise I might write up a full post explaning this with examples &c.

Comment by niplav on shortplav · 2024-09-04T09:59:20.526Z · LW · GW

Updated the link to the actual code. I computed the equilibria for the full game, and then computed the payoff per equilibrium for each player, and then took the mean for each player. I did the same but with the game with one option removed. The number in the chart is the proportion of games where removing one option from player A improved the payoff (averaged over equilibria).

If the number is >0.5, then that means that for that player, removing one option from A on average improves their payoffs. (The number of options is pre-removal). I also found this interesting, but the charts are maybe a bit misleading because often removing one option from A doesn't change the equilibria. I'll maybe generate some charts for this.

I'll perhaps also write a clearer explanation of what is happening and repost as a top-level post.

Comment by niplav on shortplav · 2024-09-03T14:11:47.666Z · LW · GW

How Often Does Taking Away Options Help?

In some game-theoretic setups, taking options away from a player improves their situation. I ran a Monte-Carlo simulation to figure out how often that is the case, generating random normal form games with payoffs in , removing a random option from the first player, and comparing the Nash equilibria found via vertex enumeration of the best response polytope (using nashpy)—the Lemke-Howson algorithm was giving me duplicate results.

Code here, largely written by Claude 3.5 Sonnet.

Comment by niplav on shortplav · 2024-08-27T20:37:24.033Z · LW · GW

I find the Thompson hack very fascinating from an agent foundations perspective. It's basically a small version of reflective stability in the context of operating systems.

I used to find compilers written in their own language kind of—…distasteful, in some way? Some of that is still present, because in reality it's just that the bootstrapping chains become very long and difficult to follow. But I think a small part of that distaste was the worry that Thompson hack-style errors occur accidentally at some point, and are just propagated through the bootstrapping chain. After thinking about this for a few seconds this was of course patently ridiculous.

But under this lens reflective stability becomes really difficult, because every replicating/successor-generating subsystem needs to be adapted to have the property of reflective stability.

E.g. corrigibility is really hard if one imagines it as a type of Thompson hack, especially under relative robustness to scale. You don't just get a basin of Thompson-hackness when writing compilers and making mistakes.

Comment by niplav on Please do not use AI to write for you · 2024-08-21T23:19:06.785Z · LW · GW

I think the splash images are >95th percentile of AI generated images in posts in beauty, especially as they still carry some of the Midjourney v1-v3 vibe, which was much more gritty and earnest (if not realistic) than the current outputs.

I really like some of the images people have used for sequences, e.g. here, here, here and here. Wikimedia has tons of creative commons images as well which I'd use if I were more into that.

Comment by niplav on Please do not use AI to write for you · 2024-08-21T21:09:21.670Z · LW · GW

Funnily enough, I feel pretty similar about AI-generated images by now. I've always struck by how people stick huge useless images on top of their blogposts, and other people want to read such articles more?? But now with AI generated images this has multiplied to a point that's just frustrating—it means more time spent scrolling, and it evokes in me the feeling of someone trying to (and failing to) set an æsthetic atmosphere for a post and then convincing me through that atmosphere, instead of providing arguments or evidence.

I have seen this being done well, especially in some of Joseph Carlsmith's posts, in which he uses a lot of images of old paintings, real photographs etc. I always thought of myself as having no taste, but now seeing other people sticking the same-flavored AI images on top of their posts makes me reconsider. (And I notice that there's a clear difference in beauty between /r/museum and /r/Art.)

Comment by niplav on Please do not use AI to write for you · 2024-08-21T20:59:49.688Z · LW · GW

You didn't even try to prompt it to write in a style you like more.

For all their virtues, fine-tuned language models are pretty bad at imitating the style of certain writers. I've instructed them to write like well-known writers, or given them medium amounts of text I've written, but they almost always fall back on that horribly dreadful HR-speak college-essayist style. Bleh.

Comment by niplav on Provably Safe AI: Worldview and Projects · 2024-08-10T12:57:37.726Z · LW · GW

That argument seems plausibly wrong to me.

Comment by niplav on How to (hopefully ethically) make money off of AGI · 2024-08-05T16:21:39.336Z · LW · GW

Take like 20% of my portfolio and throw it into some more tech/AI focused index fund. Maybe look around for something that covers some of the companies listed here on the brokerage interface that is presented to me (probably do a bit more research here)

Did you find a suitable index fund?

Comment by niplav on lemonhope's Shortform · 2024-08-02T15:22:13.678Z · LW · GW

This response is specific to AI/AI alignment, right? I wasn't "sub-tweeting" the state of AI alignment, and was more thinking of other endeavours (quantified self, paradise engineering, forecasting research).

In general, the bias towards framing can be swamped by other considerations.

Comment by niplav on RobertM's Shortform · 2024-07-31T15:21:17.941Z · LW · GW

I hadn't thought of that! Good consideration.

I checked whether the BIG-bench canary string is online in rot-13 or base64, and it's not.

Comment by niplav on lemonhope's Shortform · 2024-07-31T09:07:17.535Z · LW · GW

Strong agree. I think this is because in the rest of the world, framing is a higher status activity than filling, so independent thinkers gravitate towards the higher-status activity of framing.

Comment by niplav on Richard_Kennaway's Shortform · 2024-07-29T15:10:25.221Z · LW · GW

I don't know about signup numbers in general (the last comprehensive analysis was in 2020, when there was a clear trend)—but it definitely looks like people are still signing up for Alcor membership (six from January to March 2023).

However, in recent history, two cryonics organizations have been founded, Tomorrow.bio in Europe in 2019 (350 members signed up) and Southern Cryonics in Australia. People are being preserved, and Southern Cryonics recently suspended their first member.

Comment by niplav on Rats, Back a Candidate · 2024-07-28T14:42:55.344Z · LW · GW

I live in Europe.

Comment by niplav on You should go to ML conferences · 2024-07-24T15:04:42.133Z · LW · GW
  • Watermarks in the Sand: Impossibility of Strong Watermarking for Language Models provides nice theory for the intuition that robust watermarking is likely impossible

The link here leads nowhere, I'm afraid.

Comment by niplav on Closed Limelike Curves's Shortform · 2024-07-18T18:41:43.871Z · LW · GW

This seems true, thanks for your editing on the related pages.

Trying to collect & link the relevant pages: