Posts

CCS: Counterfactual Civilization Simulation 2024-05-02T22:54:29.773Z
The formal goal is a pointer 2024-05-01T00:27:38.856Z
Counterfactual Civilization Simulation Version -1.0 aka my application to Johannes Mayer's SPAR project 2024-03-10T10:10:51.339Z
Even if we lose, we win 2024-01-15T02:15:43.447Z

Comments

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T22:57:13.200Z · LW · GW

Oh no I mean they have the private key stored on the client side and decrypt it there.

Ideally all of this is behind a nice UI, like Signal.

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T22:55:37.478Z · LW · GW

I mean, Signal messenger has worked pretty well in my experience.

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:37:22.586Z · LW · GW

But safety research can actually disproportionally help capabilities, e.g. the development of RLHF allowed OAI to turn their weird text predictors into a very generally useful product.

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:33:09.372Z · LW · GW

I could see embedded agency being harmful though, since an actual implementation of it would be really useful for inner alignment

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:29:24.351Z · LW · GW

Some off the top of my head:

  • Outer Alignment Research (e.g. analytic moral philosophy in an attempt to extrapolate CEV) seems to be totally useless to capabilities, so we should almost definitely publish that.
  • Evals for Governance? Not sure about this since a lot of eval research helps capabilities, but if it leads to regulation that lengthens timelines, it could be net positive.

Edit: oops i didn't see tammy's comment

Comment by Pi Rogers (pi-rogers) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:18:44.187Z · LW · GW

Idea:

Have everyone who wants to share and recieve potentially exfohazardous ideas/research send out a 4096-bit RSA public key.

Then, make a clone of the alignment forum, where every time you make a post, you provide a list of the public keys of the people who you want to see the post. Then, on the client side, it encrypts the post using all of those public keys. The server only ever holds encrypted posts.

Then, users can put in their own private key to see a post. The encrypted post gets downloaded to the user's machine and is decrypted on the client side. Perhaps require users to be on open-source browsers for extra security.

Maybe also add some post-quantum thing like what Signal uses so that we don't all die when quantum computers get good enough.

Should I build this?

Is there someone else here more experienced with csec who should build this instead?

Comment by Pi Rogers (pi-rogers) on KAN: Kolmogorov-Arnold Networks · 2024-05-02T08:11:43.879Z · LW · GW

Is this a massive exfohazard? Should this have been published?

Comment by Pi Rogers (pi-rogers) on The formal goal is a pointer · 2024-05-02T08:03:53.011Z · LW · GW

Yikes, I'm not even comfortable maximizing my own CEV.

What do you think of this post by Tammy?

Where is the longer version of this? I do want to read it. :)

Well perhaps I should write it :)

Specifically, what is it about the human ancestral environment that made us irrational, and why wouldn't RL environments for AI cause the same or perhaps a different set of irrationalities?

Mostly that thing where we had a lying vs lie-detecting arms race and the liars mostly won by believing their own lies and that's how we have things like overconfidence bias and self-serving bias and a whole bunch of other biases. I think Yudkowsky and/or Hanson has written about this.

Unless we do a very stupid thing like reading the AI's thoughts and RL-punish wrongthink, this seems very unlikely to happen.

If we give the AI no reason to self-deceive, the natural instrumentally convergent incentive is to not self-deceive, so it won't self-deceive.

Again, though, I'm not super confident in this. Deep deception or similar could really screw us over.

Also, how does RL fit into QACI? Can you point me to where this is discussed?

I have no idea how Tammy plans to "train" the inner-aligned singleton on which QACI is implemented, but I think it will be closer to RL than SL in the ways that matter here.

Comment by Pi Rogers (pi-rogers) on The formal goal is a pointer · 2024-05-02T04:01:34.809Z · LW · GW

But we could have said the same thing of SBF, before the disaster happened.

I would honestly be pretty comfortable with maximizing SBF's CEV.

Please explain your thinking behind this?

TLDR: Humans can be powerful and overconfident. I think this is the main source of human evil. I also think this is unlikely to naturally be learned by RL in environments that don't incentivize irrationality (like ours did).

Sorrry if I was unclear there.

It's not, because some moral theories are not compatible with EU maximization.

I'm pretty confident that my values satisfy the VNM axioms, so those moral theories are almost definitely wrong.

And I think this uncertainty problem can be solved by forcing utility bounds.

Comment by Pi Rogers (pi-rogers) on The formal goal is a pointer · 2024-05-02T02:42:32.202Z · LW · GW

I'm 60% confident that SBF and Mao Zedong (and just about everyone) would converge to nearly the same values (which we call "human values") if they were rational enough and had good enough decision theory.

If I'm wrong, (1) is a huge problem and the only surefire way to solve it is to actually be the human whose values get extrapolated. Luckily the de-facto nominees for this position are alignment researchers, who pretty strongly self-select for having cosmopolitan altruistic values.

I think (2) is a very human problem. Due to very weird selection pressure, humans ended up really smart but also really irrational. I think most human evil is caused by a combination of overconfidence wrt our own values and lack of knowledge of things like the unilateralist's curse. An AGI (at least, one that comes from something like RL rather than being conjured in a simulation or something else weird) will probably end up with a way higher rationality:intelligence ratio, and so it will be much less likely to destroy everything we value than an empowered human. (Also 60% confident. I would not want to stake the fate of the universe on this claim)

I agree that moral uncertainty is a very hard problem, but I don't think we humans can do any better on it than an ASI. As long as we give it the right pointer, I think it will handle the rest much better than any human could. Decision theory is a bit different, since you have to put that into the utility function. Dealing with moral uncertainty is just part of expected utility maximization.

To solve (2), I think we should try to adapt something like the Hippocratic principle to work for QACI, without requiring direct reference to a human's values and beliefs (the sidestepping of which is QACI's big advantage over PreDCA). I wonder if Tammy has thought about this.

Comment by Pi Rogers (pi-rogers) on Tamsin Leake's Shortform · 2024-05-02T02:41:19.831Z · LW · GW

What about the following:

My utility function is pretty much just my own happiness (in a fun-theoretic rather than purely hedonistic sense). However, my decision theory is updateless with respect to which sentient being I ended up as, so once you factor that in, I'm a multiverse-wide realityfluid-weighted average utilitarian.

I'm not sure how correct this is, but it's possible.

Comment by Pi Rogers (pi-rogers) on The formal goal is a pointer · 2024-05-01T00:33:00.182Z · LW · GW

Edit log:

2024-04-30 19:31 CST: Footnote formatting fix and minor grammar fix.

20:40 CST: "The problem is..." --> "Alignment is..."

22:17 CST: Title changed from "All we need is a pointer" to "The formal goal is a pointer"

Comment by Pi Rogers (pi-rogers) on Tamsin Leake's Shortform · 2024-04-22T16:03:15.889Z · LW · GW

OpenAI is not evil. They are just defecting on an epistemic prisoner's dilemma.

Comment by Pi Rogers (pi-rogers) on Even if we lose, we win · 2024-03-06T01:19:17.412Z · LW · GW

Maybe some kind of simulated long-reflection type thing like QACI where "doing philosophy" basically becomes "predicting how humans would do philosophy if given lots of time and resources"

Comment by Pi Rogers (pi-rogers) on Even if we lose, we win · 2024-01-15T15:41:51.194Z · LW · GW

Yes, amount of utopiastuff across all worlds remains constant, or possibly even decreases! But I don't think amount-of-utopiastuff is the thing I want to maximize. I'd love to live in a universe that's 10% utopia and 90% paperclips! I much prefer that to a 90% chance of extinction and a 10% chance of full-utopia. It's like insurance. Expected money goes down, but expected utility goes up.

Decision theory does not imply that we get to have nice things, but (I think) it does imply that we get to hedge our insane all-or-nothing gambles for nice things, and redistribute the nice things across more worlds.

Comment by Pi Rogers (pi-rogers) on Even if we lose, we win · 2024-01-15T15:23:46.395Z · LW · GW

I think this is only true if we are giving the AI a formal goal to explicitly maximize, rather than training the AI haphazardly and giving it a clusterfuck of shards. It seems plausible that our FAI would be formal-goal aligned, but it seems like UAI would be more like us unaligned humans—a clusterfuck of shards. Formal-goal AI needs the decision theory "programmed into" its formal goal, but clusterfuck-shard AI will come up with decision theory on its own after it ascends to superintelligence and makes itself coherent. It seems likely that such a UAI would end up implementing LDT, or at least something that allows for acausal trade across the Everett branches.

Comment by Pi Rogers (pi-rogers) on Even if we lose, we win · 2024-01-15T15:10:14.213Z · LW · GW

Fixed it! Thanks! It is very confusing that half the time people talk about loss functions and the other half of the time they talk about utility functions

Comment by Pi Rogers (pi-rogers) on Diagonalization Fixed Point Exercises · 2024-01-02T07:06:29.148Z · LW · GW

Solution to 8 implemented in python using zero self-reference, where you can replace f with code for any arbitrary function on string x (escaping characters as necessary):

 f="x+'\\n'+x"
def ff(x):
return eval(f)
(lambda s : print(ff('f='+chr(34)+f+chr(34)+chr(10)+'def ff(x):'+chr(10)+chr(9)+'return eval(f)'+chr(10)+s+'('+chr(34)+s+chr(34)+')')))("(lambda s : print(ff('f='+chr(34)+f+chr(34)+chr(10)+'def ff(x):'+chr(10)+chr(9)+'return eval(f)'+chr(10)+s+'('+chr(34)+s+chr(34)+')')))")

edit: fixed spoiler tags