Posts

On Shifgrethor 2024-10-27T15:30:13.688Z
Slightly More Than You Wanted To Know: Pregnancy Length Effects 2024-10-21T01:26:02.030Z
How Big a Deal are MatMul-Free Transformers? 2024-06-27T22:28:40.888Z
Week One of Studying Transformers Architecture 2024-06-20T03:47:12.509Z
The Data Wall is Important 2024-06-09T22:54:20.070Z
LLMs seem (relatively) safe 2024-04-25T22:13:06.221Z
AI Safety Concepts Writeup: WebGPT 2023-08-11T01:35:31.196Z
Consider Multiclassing 2022-07-07T14:54:16.797Z
Alignment Risk Doesn't Require Superintelligence 2022-06-15T03:12:56.573Z
Editing Advice for LessWrong Users 2022-04-11T16:32:17.530Z

Comments

Comment by JustisMills on On Shifgrethor · 2024-11-02T02:40:08.738Z · LW · GW

(I assume you are asking why it should be rarer, not why it is rarer.)

A few reasons, including:

  • It's often given to children, where the best parenting book(s) caution against doing that too much (though of course there are tons of times it's fine)
  • It's often given to people in emotional distress, when it famously is less likely to work well

I suppose there may be lots of cases where upregulating advice would be good, and that these outweigh the common cases where downregulating it would be good. I just haven't thought of those. If you have, I'd be interested in hearing them!

Comment by JustisMills on On Shifgrethor · 2024-11-02T02:36:58.540Z · LW · GW

In the book, I don't remember and think probably it's just weird and subtle because part of the point of it is that it's alien.

As I intend it here, yeah, it would be a distinct thing. I think almost everyone knows criticism can be painful to receive and rude to give, whereas advice can feel a lot more benign.

Comment by JustisMills on On Shifgrethor · 2024-11-02T02:35:29.397Z · LW · GW

Reasonable! It strikes me as a little silly for in person conversation, but I find it fun to type and read.

Comment by JustisMills on On Shifgrethor · 2024-11-02T02:35:01.692Z · LW · GW

"Face" is pretty close, and it's cool to be reminded that that word (in this context) exists.

The main difference as I see it is that shifgrethor is narrower. At least as I propose the term be used (which is not as subtle or mysterious as in the book), it's specific to advice. You can also lose face by e.g. not responding to taunting, or something. Shifgrethor would have no opinion on that.

Comment by JustisMills on Slightly More Than You Wanted To Know: Pregnancy Length Effects · 2024-10-21T23:48:43.042Z · LW · GW

Yeah, probably. There are a few things like "meconium aspiration" that would make a literal 1:1 womb substitute insufficient to give the baby a few more weeks, and for all we know some of the 42-43 issues are direct harms of marginal gestation. But I'd be rather surprised (<10% chance) if the optimal gestation-in-artificial-womb duration were less than 41 weeks.

Comment by JustisMills on Slightly More Than You Wanted To Know: Pregnancy Length Effects · 2024-10-21T23:47:29.016Z · LW · GW

They're correlational, though the broad cohorts help - not sure what you can do beyond just canvassing an entire birth cohort and noticing differences. There are possible pitfalls like the decision to induct early being made by people with genes that predict bad outcomes? But I really don't think that's major.

Comment by JustisMills on How Big a Deal are MatMul-Free Transformers? · 2024-06-28T04:55:25.839Z · LW · GW

Yeah, you've convinced me I was a little too weak just by saying "the scaling laws are untested" - I had the same feeling of like "maybe I'm getting Eulered here, and maybe they're Eulering themselves" with the 10^23 thing.

Mostly I just kept seeing suggested articles in the mainstream-ish tech press about this "wow, no MatMul" thing, assumed it was an overhyped exaggeration/misleading, and was pleasantly surprised it was for real (as far as it goes). But I'd give it probably... 15%? Of having industrial use cases in the next few years. Which I guess is actually pretty high! Could be nice for really really huge context windows, where scaling on input token length sucks.

Comment by JustisMills on How Big a Deal are MatMul-Free Transformers? · 2024-06-28T02:40:51.705Z · LW · GW

Yeah, could cut both ways for this I think? On the one hand, if no-MatMul models really are more efficient in the long run, you could probably make custom hardware optimized for the stuff they require (e.g. lots of ternary stuff). But getting there from the ASICs currently in development would be a necessary pivot. 

Maybe the race dynamics actually help slow things down here? Since nobody wants to pivot and fall temporarily behind; money might dry up or someone else might get there before the investment pays off and you leapfrog.

But yeah, even in the medium run, as constraints start to flare up, probably ASICs are a factor in changing up architectures.

Comment by JustisMills on The Data Wall is Important · 2024-06-10T22:53:04.150Z · LW · GW

Thanks for this - helpful and concrete, and did change my mind somewhat. Of course, if it really is just 10x, in terms of orders of magnitude/hyper fast scaling we are pretty close to the wall.

Comment by JustisMills on The Data Wall is Important · 2024-06-10T22:51:54.740Z · LW · GW

Mostly just public text, I think. But I'm not sure how much more you get out of e.g. video transcripts. Maybe a lot! But it wouldn't surprise me if that was notably worse as a source.

Comment by JustisMills on The Data Wall is Important · 2024-06-10T22:50:31.463Z · LW · GW

Whoops! Thank you, fixed.

Comment by JustisMills on LLMs seem (relatively) safe · 2024-04-27T01:55:52.680Z · LW · GW

Maybe worth a slight update on how the AI alignment community would respond? Doesn't seem like any of the comments on this post are particularly aggressive. I've noticed an effect where I worry people will call me dumb when I express imperfect or gestural thoughts, but it usually doesn't happen. And if anyone's secretly thinking it, well, that's their business!

Comment by JustisMills on LLMs seem (relatively) safe · 2024-04-26T03:29:24.362Z · LW · GW

I think self-critique runs into the issues I describe in the post, though without insider information I'm not certain. Naively it seems like existing distortions would become larger with self-critique, though.

For human rating/RL, it seems true that it's possible to be sample efficient (with human brain behavior as an existence proof), but as far as I know we don't actually know how to make it sample efficient in that way, and human feedback in the moment is even more finite than human text that's just out there. So I still see that taking longer than, say, self play.

I agree that if outcome-based RL swamps initial training run datasets, then the "playing human roles" section is weaker, but is that the case now? My understanding (could easily be wrong) is that RLHF is a smaller postprocessing layer that only changes models moderately, and nowhere near the bulk of their training.

Comment by JustisMills on What do you do to deliberately practice? · 2022-06-05T02:50:18.600Z · LW · GW

I journal! It's a good way to write at least something daily, and often also feels like a good avenue for healthy introspection.

Comment by JustisMills on Increasing Demandingness in EA · 2022-04-30T17:58:57.319Z · LW · GW

I wrote a reply to this from a more-peripheral-EA perspective on the EA forum here:

https://forum.effectivealtruism.org/posts/YeudcYiArwWrg77Ng/notes-from-a-pledger

Comment by JustisMills on Austin Chen's Shortform · 2022-04-21T03:15:33.805Z · LW · GW

Thank you!

Comment by JustisMills on Editing Advice for LessWrong Users · 2022-04-12T01:48:33.109Z · LW · GW

My pleasure!

Comment by JustisMills on Editing Advice for LessWrong Users · 2022-04-12T01:44:27.977Z · LW · GW

Yeah, that critique is part of why "use more links" is among my least confident advice of the stuff in this post. I like links mostly as an alternative to nothing - if there's a term of background that ideally your readers should already know, a link is an economical way to give readers below your target audience in background knowledge a leg up. But for really central terms, yeah, better to summarize in your own words.

Comment by JustisMills on Editing Advice for LessWrong Users · 2022-04-11T18:22:59.469Z · LW · GW

Yeah, that's a good pithy summary! I often suggest replacing "this" with "this [x]".