Posts

Comments

Comment by greatBigDot on Why do we care about agency for alignment? · 2023-08-08T00:57:30.484Z · LW · GW

Even e-coli, when swimming along a sugar gradient, have an internal molecular species whose concentration roughly tracks the rate-of-change of the external sugar concentration as the bacterium swims. It's a tiny internal model of the e-coli's environment. More generally, cells often use some internal molecular species to track some external state, and update that internal representation as new information comes in.

Woah, this sounds incredibly fascinating, I've never heard of this — do you have a link to more info, or terminology to google?

Comment by greatBigDot on continue working on hard alignment! don't give up! · 2023-03-24T03:24:58.470Z · LW · GW

Let's agree that the first step towards AI alignment is to refrain from building intelligent machines that are designed to kill people.

I don't get it; why would 'refraining from designing intelligent machines to kill people' help prevent AI from killing everyone? That's a really bold and outlandish claim that I think you have to actually defend and not just tell people to agree with... Like, from my perspective, you're just assuming the hard parts of the problem don't exist, and replacing all the hard parts with an easier problem ('avoid designing AIs to kill people'). It's the hard parts of the problem that seem on track to kill us; solving the easier problem doesn't seem to help.

Comment by greatBigDot on AGI Ruin: A List of Lethalities · 2022-06-12T17:48:10.685Z · LW · GW

I don't understand; why would changing its mind about what it wants help it make more paperclips?