Posts

Daniel Dennett has died (1942-2024) 2024-04-19T16:17:04.742Z
LessWrong's (first) album: I Have Been A Good Bing 2024-04-01T07:33:45.242Z
kave's Shortform 2024-03-05T04:35:13.510Z
If you weren't such an idiot... 2024-03-02T00:01:37.314Z
New LessWrong review winner UI ("The LeastWrong" section and full-art post pages) 2024-02-28T02:42:05.801Z
On plans for a functional society 2023-12-12T00:07:46.629Z
A bet on critical periods in neural networks 2023-11-06T23:21:17.279Z
Singular learning theory and bridging from ML to brain emulations 2023-11-01T21:31:54.789Z
The Good Life in the face of the apocalypse 2023-10-16T22:40:15.200Z
How to partition teams to move fast? Debating "low-dimensional cuts" 2023-10-13T21:43:53.067Z
Navigating an ecosystem that might or might not be bad for the world 2023-09-15T23:58:00.389Z
PSA: The Sequences don't need to be read in sequence 2022-05-23T02:53:41.957Z

Comments

Comment by kave on Examples of Highly Counterfactual Discoveries? · 2024-04-26T23:45:32.228Z · LW · GW

Maybe "counterfactually robust" is an OK phrase?

Comment by kave on Take the wheel, Shoggoth! (Lesswrong is trying out changes to the frontpage algorithm) · 2024-04-26T21:02:53.475Z · LW · GW

I am sad to see you getting so downvoted. I am glad you are bringing this perspective up in the comments.

Comment by kave on My experience using financial commitments to overcome akrasia · 2024-04-26T18:20:26.743Z · LW · GW

I like comments about other users' experiences for similar reasons why I like OP. I think maybe the ideal such comment would identify itself more clearly as an experience report, but I'd rather have the report than not.

Comment by kave on Examples of Highly Counterfactual Discoveries? · 2024-04-26T18:11:39.847Z · LW · GW

What you probably mean is "completely unexpected", "surprising" or something similar

I think it means the more specific "a discovery that if it counterfactually hadn't happened, wouldn't have happened another way for a long time". I think this is roughly the "counterfactual" in "counterfactual impact", but I agree not the more widespread one.

It would be great to have a single word for this that was clearer.

Comment by kave on Elizabeth's Shortform · 2024-04-25T17:11:41.557Z · LW · GW

Enovid is also adding NO to the body, whereas humming is pulling it from the sinuses, right? (based on a quick skim of the paper).

I found a consumer FeNO-measuring device for €550. I might be interested in contributing to a replication

Comment by kave on Tamsin Leake's Shortform · 2024-04-22T20:36:32.911Z · LW · GW

(No, "you need huge profits to solve alignment" isn't a good excuse — we had nowhere near exhausted the alignment research that can be done without huge profits.)

This seems insufficiently argued; the existence of any alignment research that can be done without huge profits is not enough to establish that you don't need huge profits to solve alignment (particularly when considering things like how long timelines are even absent your intervention).

To be clear, I agree that OpenAI are doing evil by creating AI hype.

Comment by kave on CTMU insight: maybe consciousness *can* affect quantum outcomes? · 2024-04-22T18:41:55.588Z · LW · GW

Is there anything particularly quantum about this effect?

Using the simulator frame, one might think there's space to tweak:

  1. The basic physical laws
  2. The fundamental constants
  3. The "PRNG" (in an Everettian picture this looks kind of weird because its more like throwing out parts of the wavefunction to save on computation; reminds me a little of mangled worlds)

Perhaps the idea is that tweaking 1 & 2 results in worlds less interesting to the simulator?

Comment by kave on What's with all the bans recently? · 2024-04-21T00:53:11.105Z · LW · GW

I'm not seeing any active rate limits. Do you know when you observed it? It's certainly the case that an automatic rate limit could have kicked in and then, as voting changed, been removed.

Comment by kave on lukehmiles's Shortform · 2024-04-13T22:21:18.618Z · LW · GW

Good question! From the Wiki-Tag FAQ:

A good heuristic is that tag ought to have three high-quality posts, preferably written by two or more authors. 

I believe all tags have to be approved. If I were going through the morning moderation queue, I wouldn't approve an empty tag.

Comment by kave on Martín Soto's Shortform · 2024-04-11T20:40:46.307Z · LW · GW

I was trying to figure out why you believed something that seemed silly to me! I think it barely occurred to me that it's a joke.

Comment by kave on romeostevensit's Shortform · 2024-04-10T20:03:34.022Z · LW · GW

The main subcultures that I can think of where this applies are communities based around solving some problem:

  • Weight loss, especially if based around a particular diet
  • Dealing with a particular mental health problem
  • Trying to solve a particular problem in the world (e.g. explaining some mystery or finding the identity of some criminal)
Comment by kave on On Complexity Science · 2024-04-05T03:23:46.432Z · LW · GW

Any favourite examples?

Comment by kave on On Complexity Science · 2024-04-05T02:39:29.340Z · LW · GW

I think my big problem with complexity science (having bounced off it a couple of times, never having engaged with it productively) is that though some of the questions seem quite interesting, none of the answers or methods seem to have much to say.

Which is exacerbated by a tendency to imply they have answers (or at least something that is clearly going to lead to an answer)

Comment by kave on Thomas Kwa's Shortform · 2024-04-04T19:48:03.331Z · LW · GW

I would like to read it! Satire is sometimes helpful for me to get a perspective shift

Comment by kave on What's with all the bans recently? · 2024-04-04T07:02:04.453Z · LW · GW

To answer, for now, just one piece of this post:

We're currently experimenting with a rule that flags users who've received several downvotes from "senior" users (I believe 5 downvotes from users with above 1,000 karma) on comments that are already net-negative (I believe that were posted in the last year).

We're currently in the manual review phase, so users are being flagged and then users are having the rate limit applied if it seems reasonable. For what it's worth, I don't think this rule has an amazing track record so far, but all the cases in the "rate limit wave" were reviewed by me and Habryka and he decided to apply a limit in those cases.

(We applied some rate limit in 60% of the cases of users who got flagged by the rule).

People who get manually rate-limited don't have an explanation visible when trying to comment (unlike users who are limited by an automatic rule, I think).

We have explained this to users that reached out (in fact this answer is adapted from one such conversation), but I do think we plausibly should have set up infrastructure to explain these new rate limits.

Comment by kave on Open Thread Spring 2024 · 2024-04-03T20:23:25.069Z · LW · GW

Hello and welcome to the site! I'm glad you're saying hello despite having been too shy :-)

Do let us know in this thread or in the intercom in the bottom right if you run into any problems.

Comment by kave on My PhD thesis: Algorithmic Bayesian Epistemology · 2024-04-03T20:21:21.543Z · LW · GW

Curated.

Using Bayes-type epistemology is a core LessWrong topic, and I think this represents a bunch of progress on that front (whether the results are already real-world-ready or just real-world-inspired). I have only engaged with small parts of the thesis, but those parts seem pretty exciting; so far, I particularly like knowing about quasi-arithmetic pooling. It feels like I've become less confused about something that I didn't know I was confused about — the connection between the character of the proper scoring rule and the right ways to aggregate those probabilities.

I also appreciate Eric's work making blogposts explaining more of his thoughts in a friendly way. Hope to see a few more distillations come out of this thesis!

Comment by kave on LessWrong's (first) album: I Have Been A Good Bing · 2024-04-03T07:59:59.539Z · LW · GW

Much sweat and some tears were spent on trying to get something like that working, but the Shoggoths are fickle

Comment by kave on LessWrong's (first) album: I Have Been A Good Bing · 2024-04-01T23:25:25.831Z · LW · GW

Thanks so much for the fix!

Comment by kave on Daniel Kahneman has died · 2024-03-28T21:14:38.557Z · LW · GW

(I assume you mean the story with him and the SS soldier; I think a couple of people got confused and thought you were referring to the fact Kahneman had died)

Comment by kave on Leading The Parade · 2024-03-24T18:45:59.500Z · LW · GW

Say more / references?

Comment by kave on Vernor Vinge, who coined the term "Technological Singularity", dies at 79 · 2024-03-22T19:47:26.149Z · LW · GW

From a message I wrote to a friend once that seems a little relevant

[H]ow should you act when you’re inside someone’s OODA loop? I was thinking about how like Wikipedia/tab explosions are sort of inside my ooda loop. But sometimes I can be more of an active reader who is navigating the concepts being exposed to me as I choose, and the process becomes like a magic genie or butler who is doing interpretative labour and conjuring up new scenes following my fickle interest.

So it seems like one thing that the person with the smaller loop can do is interpretative labour, and spend the faster cycles on self-legibilising.

Comment by kave on Increasing IQ by 10 Points is Possible · 2024-03-21T16:30:11.805Z · LW · GW

Yep, the question is definitely about how far it transfers.

Comment by kave on Richard Ngo's Shortform · 2024-03-20T21:56:29.195Z · LW · GW

increase the agent's expected future value

I wonder if there's a loopiness here is which breaks the setup (the expectation I'm guessing is relative to the prediction markets probabilities? Though it seems like the market is over sensory experiences but the values are over world states in general, so maybe I'm missing something). But it seems like if I take an action and move the market at the same time, I might be able to extract a bunch of extra money and acquire outsize control.

  1. Bidding to control the agent's actions for the next N timesteps

This seems like it's wasteful relative to contributing to a pool that bids on action A (or short-term policy P). I guess coordination is hard if you're just contributing to the pool though, and all connects to the merging process you describe.

Comment by kave on Increasing IQ by 10 Points is Possible · 2024-03-20T19:15:04.992Z · LW · GW

I mean when I journal I come up with little exercises to improve areas of my life. I imagine that people in your cohort might do similarly, and given that they signed up to improve their IQ, that might include things adjacent to the tasks of the IQ test.

And I don't think general meditation should count as training, but specific meditations could (e.g. if you are training doing mental visualisations and the task involves mental rotations).

I'm not trying to say that there are definitely cross-training effects, just that these seem like the kinds of thing which are somewhat more likely (than, say, supplements) to create fairly narrow improvements close to the test.

Comment by kave on Increasing IQ by 10 Points is Possible · 2024-03-20T01:35:11.292Z · LW · GW

And I can make people think “out of the box” (e.g. via specific games, specific “supplements”, specific meditations)

And prod people to think about how they can improve in whatever areas they want (e.g. via journaling, talking, and meditating)

 

Ah, these two have made me more concerned about training effects: especially the games, but also the meditations and journaling.

It seems pretty plausible certain games could basically train the same skills as the IQ test.

Comment by kave on How do you feel about LessWrong these days? [Open feedback thread] · 2024-03-19T17:56:27.197Z · LW · GW

I think this a real problem (tho I think it's more fundamental than your hypothesis would suggest; we could check commenting behaviour in the 2000s as a comparison).

We have some explorations underway addressing related issues (like maybe the frontpage should be more recommender-y and show you good old posts, while the All Posts page is used for people who care a lot about recency). I don't think we've concretely considered stuff that would show you good old posts with new comments, but that might well be worth exploring.

Comment by kave on 'Empiricism!' as Anti-Epistemology · 2024-03-18T22:18:11.478Z · LW · GW

My quick guess is that people don't agree about what constitutes a (relevant) flaw. (And there are lots of irrelevant flaws so you can't just check for the existence of any flaws at all).

I think if people could agree, the authorial incentives would follow. I'm fairly sympathetic to the idea that readers aren't incentivised to correctly agree on what consitutes a flaw.

Comment by kave on New social credit formalizations · 2024-03-15T17:30:49.225Z · LW · GW

I notice that the formalised owings have professionals dedicated (in part) to making sure people line up their behaviour with the formalised owings. I wonder if other social credit formalizations would also need this

Comment by kave on To the average human, controlled AI is just as lethal as 'misaligned' AI · 2024-03-14T21:51:24.853Z · LW · GW

I recognised the allusion but also disliked the title.

Comment by kave on kave's Shortform · 2024-03-14T21:48:15.551Z · LW · GW

Sometimes running to stand still is the right thing to do

It's nice when good stuff piles up into even more good stuff, but sometimes it doesn't:

  • Sometimes people are worried that they will habituate to caffeine and lose any benefit from taking it.
  • Most efforts to lose weight are only temporarily successful (unless using medicine or surgery).
  • The hedonic treadmill model claims it’s hard to become durably happier.
  • Productivity hacks tend to stop working.

These things are like Alice’s red queen’s race: always running to stay in the same place. But I think there’s a pretty big difference between running that keeps you exactly where you would have been if you hadn’t bothered, and running that either moves you a little way and then stops, or running that stops you moving in one direction.

I’m not sure what we should call such things, but one idea is hamster wheels for things that make no difference, bungee runs for things that let you move in a direction a bit but you have to keep running to stay there, and backwards escalators for things where you’re fighting to stay in the same place rather than moving in a direction (named for the grand international pastime of running down rising escalators).

I don't know which kind of thing is most common, but I like being able to ask which dynamic is at play. For example, I wonder if weight loss efforts are often more like backwards escalators than hamster wheels. People tend to get fatter as they get older. Maybe people who are trying (but failing) to lose weight are gaining weight more slowly than similar people who aren’t trying to do so?

Or my guess is that most people will have more energy than baseline if they take caffeine every day, even though any given dose will have less of an effect than taking the same amount of caffeine while being caffeine-naive, so they've bungee ran (done a bungee run?) a little way forward and that's as far as they'll go.

I am currently considering whether productivity hacks, which I’ve sworn off, are worth doing even though they only last for a little while. The extra, but finite, productivity could be worth it. (I think this would count as another bungee run).

I'd be interested to hear examples that fit within or break this taxonomy.

Comment by kave on 'Empiricism!' as Anti-Epistemology · 2024-03-14T18:59:52.433Z · LW · GW

I sometimes like things being said in a long way. Mostly that's just because it helps me stew on the ideas and look at them from different angles. But also, specifically, I liked the engagement with a bunch of epistemological intuitions and figuring out what can be recovered from them. I like in particular connecting the "trend continues" trend to the redoubtable "electron will weight the same tomorrow" intuition.

(I realise you didn't claim there was nothing else in the dialogue, just not enough to justify the length)

Comment by kave on 'Empiricism!' as Anti-Epistemology · 2024-03-14T05:24:21.660Z · LW · GW

Twitter

Comment by kave on Some (problematic) aesthetics of what constitutes good work in academia · 2024-03-13T00:54:19.024Z · LW · GW

How many people read your post is probably meaningful to you, and karma affects that a lot.

I say this because I certainly care about how many people read which posts, so it's kind of sad when karma doesn't track value in the post (though of course brevity and ease of reading are also important and valuable).

Comment by kave on LawrenceC's Shortform · 2024-03-12T23:05:24.106Z · LW · GW
  • -- Some beliefs about AI Scaling Labs that I'm redacting on LW --

Is the reason for the redaction also private?

Comment by kave on A guide to Iterated Amplification & Debate · 2024-03-12T18:40:25.266Z · LW · GW

Hm, I'm not seeing the issue. Could you screenshot it?

Comment by kave on One-shot strategy games? · 2024-03-11T21:33:28.751Z · LW · GW

There are separate random number generators for most things in Slay the Spire.

Specifically, this piece of code is executed at the start of a run:

public static void generateSeeds() {
    logger.info("Generating seeds: " + Settings.seed);
    monsterRng = new Random(Settings.seed);
    eventRng = new Random(Settings.seed);
    merchantRng = new Random(Settings.seed);
    cardRng = new Random(Settings.seed);
    treasureRng = new Random(Settings.seed);
    relicRng = new Random(Settings.seed);
    potionRng = new Random(Settings.seed);
    // The following rngs are actually re-initialized each floor:
    monsterHpRng = new Random(Settings.seed);
    aiRng = new Random(Settings.seed);
    shuffleRng = new Random(Settings.seed);
    cardRandomRng = new Random(Settings.seed);
    miscRng = new Random(Settings.seed);
}

From https://forgottenarbiter.github.io/Correlated-Randomness/ (which does point out there is some unfortunate correlated randomness, though I think there are mods that fix that)

Comment by kave on One-shot strategy games? · 2024-03-11T02:16:49.606Z · LW · GW

Dungeon Crawl Stone Soup

Comment by kave on Jimrandomh's Shortform · 2024-03-08T18:59:16.041Z · LW · GW

I had previously guessed air movement made me feel better because my body expected air movement (i.e. some kind of biophilic effect). But this explanation seems more likely in retrospect! I'm not quite sure how to run the calculation using the diffusivity coefficient to spot check this, though.

Comment by kave on Vote on Anthropic Topics to Discuss · 2024-03-08T18:13:42.721Z · LW · GW

That's a cool idea! Currently this is just a restyling of a comment thread using the normal react system; I think if we decided to invest more into it to make it a real feature it would be kind of cool to build stuff like that. Could also be cool to make it play with viewpoints somehow, though the philosophies are a little different. These polls are built around people publically stating their opinions, whereas viewpoints is anonymous(?).

Comment by kave on AI #54: Clauding Along · 2024-03-07T21:15:17.694Z · LW · GW

I think it's a fairly narrow set of assumptions where £20 for £7 is worse than £10 for free, but £20 for £8 is better than £10 for £1.

Comment by kave on Using axis lines for good or evil · 2024-03-07T18:54:45.601Z · LW · GW

When I looked at your proposed GDP-Time chart, I felt I was more inclined to treat the year as quantitative and the amounts as categorical. Though I don't know how that would actually play out if I were trying to make use of it in anger.

Comment by kave on Vote on Anthropic Topics to Discuss · 2024-03-07T01:10:41.017Z · LW · GW

Good idea! I made some here

Comment by kave on Vote on Anthropic Topics to Discuss · 2024-03-07T01:10:14.370Z · LW · GW
Comment by kave on Vote on Anthropic Topics to Discuss · 2024-03-06T23:12:15.385Z · LW · GW

I assign >10% that Anthropic will at some point pause development for at least a year as a result of safety evaluations.

I'm at >10% including "completely halt" and at <10% not including that.

Comment by kave on Vote on Anthropic Topics to Discuss · 2024-03-06T22:54:49.734Z · LW · GW

Regardless of whether or not Claude 3 was significant progress over GPT-4, it worsened race dynamics to market it as being so.

Comment by kave on Anthropic release Claude 3, claims >GPT-4 Performance · 2024-03-06T19:28:10.526Z · LW · GW

I hear you sometimes share dual-use (or plain capabilities?) ideas with Anthropic. If that's true, does this change your policy?

Comment by kave on Fabien's Shortform · 2024-03-05T19:05:32.890Z · LW · GW

only because people did not argue with Einstein by commenting on how crazy the theory was

Did Einstein's theory seem crazy to people at the time?

Comment by kave on kave's Shortform · 2024-03-05T04:35:32.781Z · LW · GW

Suppose you are a government that thinks some policy will be good for your populace in aggregate over the long-term (that is, it's a Kaldor-Hicks improvement). For example, perhaps some tax reform you're excited about.

But this reform (we assume) is quite unpopular with a few people who would suffer concentrated losses. You're tempted to include in the policy a big cash transfer that makes those people happy (that is, making it closer to Pareto optimal). But you're worried about levering up too much on your guess that this is a good policy.

Here's one thing you can do. You auction securities (that is, tradeable assets) that pay off as follows: if you raise $X through auctioning off the securities, you are committed to the policy (including the big cash transfer) and the security converts into one that tracks something about how well your populace is doing (like a share of an index fund or something). If you raise less than that, the owner of the security gets back the money they spent on the asset.

Ignoring some annoying details like the operational costs of this scheme or the foregone interest while waiting for the security to activate a branch of the conditional, the value of that security (which should be equal to the price you paid for it, if you didn't capture any surplus) is just the value of the security it converts to.

(Solve  +  = )

So this scheme lets you raise the cash for your policy under exactly the conditions when the auction "thinks" the value of the security increases sufficiently. Which is kind of neat.

Comment by kave on If you weren't such an idiot... · 2024-03-04T23:29:50.003Z · LW · GW

I don't have strong takes on the "idiot" stylistic choice; introspectively it feels fun and it takes the bite out of my self-flagellations when I notice leaving value on the floor. I won't claim to have any empirical support that this actually works better or worse for me.

I mostly like this post for the advice. Concretely, after reading it:

  • I've bought copies of exercise equipment I like for my parents' house so that when I am exercising relatively regularly, I don't break a streak by visiting them.
  • I own two laptops, two kindles, have multiple chargers (some of which live in travel bags so I always have them)
  • I often invoke this when deciding whether to try an experimental purchase that might make my life better, but is a little pricey.
  • Rebought multi-colour pens, which tend to be useful for me for a stretch, and I don't consider rebuying them when they might be useful again.
  • Bought an external battery.

That collection feels like it was probably net-positive.

Some of the things that Mark suggests have been really good for me but I already did them (have a password manager, bright rooms).  There are some I disagree with, so haven't (re)tried after reading it (drink lots of water, summarize things you've read, reliably sleep 6-9 hours a night (on the margin, for me)). And then there are some that seem likely good that I haven't tried.