A Meditative Experience 2021-12-03T17:58:39.462Z


Comment by Yonatan Cale (yonatan-cale-1) on The Plan · 2021-12-29T16:12:30.099Z · LW · GW

Hypothesis regarding your confusion about agency:

Describing humans using a "utility function" or through "goals" is wrong.

Humans are a bunch of habits (like CFAR TAPs) which have some correlation with working towards goals, but this is more of an imperfect rationalization than a reasonable/natural way to describe the situation.

Also yes, we have some part that thinks in goals, but it has a very limited effect on anything (like actions) compared to what we'd naturally think.

Credit to a friend

[I have no idea what I'm talking about, feel free to ignore if this doesn't resonate of course, seemed worth a comment]

Comment by Yonatan Cale (yonatan-cale-1) on How To Write Quickly While Maintaining Epistemic Rigor · 2021-12-27T11:20:00.558Z · LW · GW

This has influenced how I write things and how I give disclaimers (I read it a few months ago)


Comment by Yonatan Cale (yonatan-cale-1) on Attempted Gears Analysis of AGI Intervention Discussion With Eliezer · 2021-11-22T22:34:20.423Z · LW · GW

Some people here inspire me to make predictions ;) So here's my attempt:

My guess, mainly based on this image (linked from the post):


Is that he'd say it's a sub category of "getting models to output things based only on their training data, while treating them as a black box and still assuming unexpected outputs will happen sometimes", as well as "this might work well for training, but obviously not for an AGI" and "if we're going to talk about limiting a model's output, Redwood Research is more of a way to go" and perhaps "this will just advance AI faster"

Comment by Yonatan Cale (yonatan-cale-1) on Attempted Gears Analysis of AGI Intervention Discussion With Eliezer · 2021-11-22T22:33:55.153Z · LW · GW

[AI risk question, not sure where to ask]

Hey, could you (or someone) help me understand how useful this would be? (Or, what would Yudkowsky say about it?)

I'm asking because this might be something that I, or someone that I know, could do

Comment by Yonatan Cale (yonatan-cale-1) on Shoulder Advisors 101 · 2021-11-07T23:31:46.871Z · LW · GW

@hnykda this is why I have a vmDaniel.  You see, it's even Duncan Approved