Posts

What AI safety researchers can learn from Mahatma Gandhi 2024-11-08T19:49:14.397Z
Is OpenAI net negative for AI Safety? 2024-11-02T16:18:02.859Z
Thirty random thoughts about AI alignment 2024-09-15T16:24:10.572Z
Disproving and partially fixing a fully homomorphic encryption scheme with perfect secrecy 2024-05-26T14:56:17.233Z
Planning to build a cryptographic box with perfect secrecy 2023-12-31T09:31:47.941Z

Comments

Comment by Lysandre Terrisse on Interlude for Behavioral Economics · 2024-10-27T10:14:37.058Z · LW · GW

People from “primitive” societies give more than people from more developed societies, and the more primitive the society, the stronger the effect.

 

I am not sure what you intended to say here, but the word "primitive" definitely looks like a red flag. As I don't think I am the only one to believe this, I would ask you to please change the wording or delete this sentence.

Comment by Lysandre Terrisse on Planning to build a cryptographic box with perfect secrecy · 2024-01-01T19:00:13.062Z · LW · GW

Thanks for the comment! I don't think you are interpreting this post wrong.

Why would we want to run it if we can't see what it does?

I don't think that running a superintelligence is ever useful. What I argue is that running it inside a cryptographic box is better than running it outside of the box. The goal of this project is that, if a superintelligence is run, then it doesn't automatically destroy the world. The goal isn't to run a superintelligence to do useful tasks.

Like maybe for , it would not be enough to persuade a person to do something that would unbox the AI (but it might).

Personally, I think that making it output  bits of information is too much. If the superintelligence were to output an ASCII text (which is very inefficient in order to unbox itself), then  bits would correspond to 8192 characters, which seems way more than enough in order to convince someone to unbox it. If the decrypted bits were thoroughly examined by people outside of the box, then I expect that  would be too much. It would be interesting to know how much information we are able to communicate with that many bits (it makes me think about the Arecibo message, which uses 1679 bits).

Since I think that running a superintelligence is already too dangerous even without intending to decrypt anything, I wouldn't approve a team that tries to build a superintelligence with . However, I think that planning  is better than not trying to use a cryptographic box at all.

I don't think having a formal proof should be an objective in and of itself.

I think the objective should be to mitigate the risk of extinction by AI. Developing an unconditional proof of safety seems to be one way to reduce this risk, and this is why I decided to work on cryptographic boxes. However, we shouldn't rely only on trying to develop unconditional proofs, since these seem possible only when talking about control (how to prevent a superintelligence from achieving its bad objectives), but they seem nearly impossible to do when talking about alignment (how to give good objectives to a superintelligence). Moreover, working on alignment seems more promising than working on control in the cases where we have a lot of time until the first superintelligence.

If we don't allow any output bits to come out of the ASI, then how do we know if it worked?

I think that, if the box were to have holes, then we would realize it pretty soon from the fact that the superintelligence would dramatically change the world around us, and probably make Earth uninhabitable in this process.

Comment by Lysandre Terrisse on Planning to build a cryptographic box with perfect secrecy · 2024-01-01T18:29:33.965Z · LW · GW

Thank you! Personally, I think that, if a layperson were trying to help me, they could do it by trying to find flaws in the plan. I already mentioned that the One-Time Pad used to fail during WWII in an unexpected way, despite the fact that it had a proof of perfect secrecy. If someone were to find a flaw in the plan, it would help me a lot (although it would also prove that my goal is impossible).