Posts

Epistemic states as a potential benign prior 2024-08-31T18:26:14.093Z
Please stop publishing ideas/insights/research about AI 2024-05-02T14:54:15.691Z
How "Pause AI" advocacy could be net harmful 2023-12-26T16:19:20.724Z
OpenAI, DeepMind, Anthropic, etc. should shut down. 2023-12-17T20:01:22.332Z
Alignment work in anomalous worlds 2023-12-16T19:34:26.202Z
Some biases and selection effects in AI risk discourse 2023-12-12T17:55:15.759Z
How LDT helps reduce the AI arms race 2023-12-10T16:21:44.409Z
We're all in this together 2023-12-05T13:57:46.270Z
So you want to save the world? An account in paladinhood 2023-11-22T17:40:33.048Z
an Evangelion dialogue explaining the QACI alignment plan 2023-06-10T03:28:47.096Z
formalizing the QACI alignment formal-goal 2023-06-10T03:28:29.541Z
Orthogonal's Formal-Goal Alignment theory of change 2023-05-05T22:36:14.883Z
Orthogonal: A new agent foundations alignment organization 2023-04-19T20:17:14.174Z
continue working on hard alignment! don't give up! 2023-03-24T00:14:35.607Z
the QACI alignment plan: table of contents 2023-03-21T20:22:00.865Z
your terminal values are complex and not objective 2023-03-13T13:34:01.195Z
the quantum amplitude argument against ethics deduplication 2023-03-12T13:02:31.876Z
QACI: the problem of blob location, causality, and counterfactuals 2023-03-05T14:06:09.372Z
state of my alignment research, and what needs work 2023-03-03T10:28:34.225Z
Hello, Elua. 2023-02-23T05:19:07.246Z
a narrative explanation of the QACI alignment plan 2023-02-15T03:28:34.710Z
so you think you're not qualified to do technical alignment research? 2023-02-07T01:54:51.952Z
formal alignment: what it is, and some proposals 2023-01-29T11:32:33.239Z
to me, it's instrumentality that is alienating 2023-01-27T18:27:19.062Z
one-shot AI, delegating embedded agency and decision theory, and one-shot QACI 2022-12-23T04:40:31.880Z
our deepest wishes 2022-12-20T00:23:32.892Z
all claw, no world — and other thoughts on the universal distribution 2022-12-14T18:55:06.286Z
a rough sketch of formal aligned AI using QACI 2022-12-11T23:40:37.536Z
Tamsin Leake's Shortform 2022-11-20T18:25:17.811Z
logical vs indexical dignity 2022-11-19T12:43:03.033Z
generalized wireheading 2022-11-18T20:18:53.664Z
fully aligned singleton as a solution to everything 2022-11-12T18:19:59.378Z
a casual intro to AI doom and alignment 2022-11-01T16:38:31.230Z
publishing alignment research and exfohazards 2022-10-31T18:02:14.047Z
love, not competition 2022-10-30T19:44:46.030Z
QACI: question-answer counterfactual intervals 2022-10-24T13:08:54.457Z
some simulation hypotheses 2022-10-12T13:34:51.780Z
confusion about alignment requirements 2022-10-06T10:32:49.779Z
my current outlook on AI risk mitigation 2022-10-03T20:06:48.995Z
existential self-determination 2022-09-27T16:08:46.997Z
ordering capability thresholds 2022-09-16T16:36:59.172Z
ethics and anthropics of homomorphically encrypted computations 2022-09-09T10:49:08.316Z
program searches 2022-09-05T20:04:17.916Z
everything is okay 2022-08-23T09:20:33.250Z
PreDCA: vanessa kosoy's alignment protocol 2022-08-20T10:03:10.701Z
goal-program bricks 2022-08-13T10:08:41.532Z
the Insulated Goal-Program idea 2022-08-13T09:57:47.251Z
The Peerless 2022-04-13T01:07:08.767Z
Nantes, France – ACX Meetups Everywhere 2021 2021-08-23T08:46:08.899Z

Comments

Comment by Tamsin Leake (carado-1) on The case for more Alignment Target Analysis (ATA) · 2024-09-20T06:44:57.578Z · LW · GW

Hi !

ATA is extremely neglected. The field of ATA is at a very early stage, and currently there does not exist any research project dedicated to ATA. The present post argues that this lack of progress is dangerous and that this neglect is a serious mistake.

I agree it's neglected, but there is in fact at least one researh project dedicated to at least designing alignment targets: the part of the formal alignment agenda dedicated to formal outer alignment, which is the design of math problems to which solutions would be world-saving. Our notable attempts at this are QACI and ESP (there was also some work on a QACI2, but it predates (and in-my-opinion is superceded by) ESP).

Those try to implement CEV in math. They only work for doing CEV of a single person or small group, but that's fine: just do CEV of {a single person or small group} which values all of humanity/moral-patients/whatever getting their values satisfied instead of just that group's values. If you want humanity's values to be satisfied, then "satisfying humanity's values" is not opposite to "satisfy your own values", it's merely the outcome of "satisfy your own values".

Comment by Tamsin Leake (carado-1) on Being nicer than Clippy · 2024-09-20T06:38:21.343Z · LW · GW
  1. I wonder how much of those seemingly idealistic people retained power when it was available because they were indeed only pretending to be idealistic. Assuming one is actually initially idealistic but then gets corrupted by having power in some way, one thing someone can do in CEV that you can't do in real life is reuse the CEV process to come up with even better CEV processes which will be even more likely to retain/recover their just-before-launching-CEV values. Yes, many people would mess this up or fail in some other way in CEV; but we only need one person or group who we'd be somewhat confident would do alright in CEV. Plausibly there are at least a few eg MIRIers who would satisfy this. Importantly, to me, this reduces outer alignment to "find someone smart and reasonable and likely to have good goal-content integrity", which is a matter of social & psychology that seems to be much smaller than the initial full problem of formal outer alignment / alignment target design.

  2. One of the main reasons to do CEV is because we're gonna die of AI soon, and CEV is a way to have infinite time to solve the necessary problems. Another is that even if we don't die of AI, we get eaten by various moloch instead of being able to safely solve the necessary problems at whatever pace is necessary.

Comment by Tamsin Leake (carado-1) on Being nicer than Clippy · 2024-09-20T06:23:03.629Z · LW · GW

the main arguments for the programmers including all of [current?] humanity in the CEV "extrapolation base" […] apply symmetrically to AIs-we're-sharing-the-world-with at the time

I think timeless values might possibly help resolve this; if some {AIs that are around at the time} are moral patients, then sure, just like other moral patients around they should get a fair share of the future.

If an AI grabs more resources than is fair, you do the exact same thing as if a human grabs more resources than is fair: satisfy the values of moral patients (including ones who are no longer around) not weighed by how much leverage they current have over the future, but how much leverage they would have over the future if things had gone more fairly/if abuse/powergrab/etc wasn't the kind of thing that gets your more control of the future.

"Sorry clippy, we do want you to get some paperclips, we just don't want you to get as many paperclips as you could if you could murder/brainhack/etc all humans, because that doesn't seem to be a very fair way to allocate the future." — and in the same breath, "Sorry Putin, we do want you to get some of whatever-intrinsic-values-you're-trying-to-satisfy, we just don't want you to get as much as ruthlessly ruling Russia can get you, because that doesn't seem to be a very fair way to allocate the future."

And this can apply regardless of how much of clippy already exists by the time you're doing CEV.

Comment by Tamsin Leake (carado-1) on Being nicer than Clippy · 2024-09-20T06:07:10.703Z · LW · GW

trying to solve morality by themselves

It doesn't have to be by themselves; they can defer to others inside CEV, or come up with better schemes that their initial CEV inside CEV and then defer to that. Whatever other solutions than "solve everything on your own inside CEV" might exist, they can figure those out and defer to them from inside CEV. At least that's the case in my own attempts at implementing CEV in math (eg QACI).

Comment by Tamsin Leake (carado-1) on Lucius Bushnaq's Shortform · 2024-09-18T20:20:17.008Z · LW · GW

Seems really wonky and like there could be a lot of things that could go wrong in hard-to-predict ways, but I guess I sorta get the idea.

I guess one of the main things I'm worried about is that it seems to require that we either:

  • Be really good at timing when we pause it to look at its internals, such that we look at the internals after it's had long enough to think about things that there are indeed such representations, but not long enough that it started optimizing really hard such that we either {die before we get to look at the internals} or {the internals are deceptively engineered to brainhack whoever would look at them}. If such a time interval even occurs for any amount of time at all.
  • Have an AI that is powerful enough to have powerful internals-about-QACI to look at, but corrigible enough that this power is not being used to do instrumentally convergent stuff like eat the world in order to have more resources with which to reason.

Current AIs are not representative of what dealing with powerful optimizers is like; when we'll start getting powerful optimizers, they won't sit around long enough for us to look at them and ponder, they'll just quickly eat us.

Comment by Tamsin Leake (carado-1) on Lucius Bushnaq's Shortform · 2024-09-18T15:25:29.881Z · LW · GW

So the formalized concept is Get_Simplest_Concept_Which_Can_Be_Informally_Described_As("QACI is an outer alignment scheme consisting of…") ? Is an informal definition written in english?

It seems like "natural latent" here just means "simple (in some simplicity prior)". If I read the first line of your post as:

Has anyone thought about QACI could be located in some simplicity prior, by searching the prior for concepts matching(??in some way??) some informal description in english?

It sure sounds like I should read the two posts you linked (perhaps especially this one), despite how hard I keep bouncing off of the natural latents idea. I'll give that a try.

Comment by Tamsin Leake (carado-1) on Lucius Bushnaq's Shortform · 2024-09-18T10:22:41.357Z · LW · GW

To me kinda the whole point of QACI is that it tries to actually be fully formalized. Informal definitions seem very much not robust to when superintelligences think about them; fully formalized definitions are the only thing I know of that keep meaning the same thing regardless of what kind of AI looks at it or with what kind of ontology.

I don't really get the whole natural latents ontology at all, and mostly expect it to be too weak for us to be able to get reflectively stable goal-content integrity even as the AI becomes vastly superintelligent. If definitions are informal, that feels to me like degrees of freedom in which an ASI can just pick whichever values make its job easiest.

Perhaps something like this allows use to use current, non-vastly-superintelligent AIs to help design a formalized version of QACI or ESP which itself is robust enough to be passed to superintelligent optimizers; but my response to this is usually "have you tried first formalizing CEV/QACI/ESP by hand?" because it feels like we've barely tried and like reasonable progress can be made on it that way.

Perhaps there are some cleverer schemes where the superintelligent optimizer is pointed at the weaker current-tech-level AI, itself pointed informally at QACI, and we tell the superintelligent optimizer "do what this guy says"; but that seems like it either leaves too many degrees of freedom to the superintelligent optimizer again, or it requires solving corrigibility (the superintelligent optimizer is corrigibly assisting the weaker AI) at which point why not just point the corrigibility at the human directly and ignore QACI altogether, at least to begin with.

Comment by Tamsin Leake (carado-1) on Epistemic states as a potential benign prior · 2024-09-01T06:53:43.453Z · LW · GW

The knightian in IB is related to limits of what hypotheses you can possibly find/write down, not - if i understand so far - about an adversary. The adversary stuff is afaict mostly to make proofs work.

I don't think this makes a difference here? If you say "what's the best not-blacklisted-by-any-knightian-hypothesis action", then it doesn't really matter if you're thinking of your knightian hypotheses as adversaries trying to screw you over by blacklisting actions that are fine, or if you're thinking of your knightian hypotheses as a more abstract worst-case-scenario. In both cases, for any reasonable action, there's probly a knightian hypothesis which blacklists it.

Regardless of whether you think of it as "because adversaries" or just "because we're cautious", knightian uncertainty works the same way. The issue is fundamental to doing maximin over knightian hypotheses.

Comment by Tamsin Leake (carado-1) on Pi Rogers's Shortform · 2024-08-26T23:13:34.468Z · LW · GW

This is indeed a meaningful distinction! I'd phrase it as:

  • Values about what the entire cosmos should be like
  • Values about what kind of places one wants one's (future) selves to inhabit (eg, in an internet-like upload-utopia, "what servers does one want to hang out on")

"Global" and "local" is not the worst nomenclature. Maybe "global" vs "personal" values? I dunno.

my best idea is to call the former "global preferences" and the latter "local preferences", but that clashes with the pre-existing notion of locality of preferences as the quality of terminally caring more about people/objects closer to you in spacetime

I mean, it's not unrelated! One can view a utility function with both kinds of values as a combination of two utility functions: the part that only cares about the state of the entire cosmos and the part that only cares about what's around them (see also "locally-caring agents").

(One might be tempted to say "consequentialist" vs "experiential", but I don't think that's right — one can still value contact with reality in their personal/local values.)

Comment by Tamsin Leake (carado-1) on Wei Dai's Shortform · 2024-08-26T22:48:58.656Z · LW · GW

That is, in fact, a helpful elaboration! When you said

Most people who "work on AI alignment" don't even think that thinking is a thing.

my leading hypotheses for what you could mean were:

  • Using thought, as a tool, has not occured to most such people
  • Most such people have no concept whatsoever of cognition as being a thing, the way people in the year 1000 had no concept whatsoever of javascript being a thing.

Now, instead, my leading hypothesis is that you mean:

  • Most such people are failing to notice that there's an important process, called "thinking", which humans do but LLMs "basically" don't do.

This is a bunch more precise! For one, it mentions AIs at all.

Comment by Tamsin Leake (carado-1) on quila's Shortform · 2024-07-27T11:09:36.451Z · LW · GW

To be more precise: extrapolated over time, for any undesired selection process or other problem of that kind, either the problem is large enough that it gets exarcerbated over time so much that it eats everything — and then that's just extinction, but slower — or it's not large enough to win out and aligned superintelligence(s) + coordinated human action is enough to stamp it out in the long run, which means they won't be an issue for almost all of the future.

It seems like for a problem to be just large enough that coordination doesn't stamp it away, but also it doesn't eat everything, would be a very fragile equilibrium, and I think that's pretty unlikely.

Comment by Tamsin Leake (carado-1) on quila's Shortform · 2024-06-01T11:16:01.434Z · LW · GW

single-use

Considering how loog it took me to get that by this you mean "not dual-use", I expect some others just won't get it.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-06-01T09:00:07.205Z · LW · GW

Some people who are very concerned about suffering might be considering building an unaligned AI that kills everyone just to avoid the risk of an AI takeover by an AI aligned to values which want some people to suffer.

Let this be me being on the record saying: I believe the probability of {alignment to values that strongly diswant suffering for all moral patients} is high enough, and the probability of {alignment to values that want some moral patients to suffer} is low enough, that this action is not worth it.

I think this applies to approximately anyone who would read this post, including heads of major labs in case they happen to read this post and in case they're pursuing the startegy of killing everyone to reduce S-risk.

See also: how acausal trade helps in 1, 2, but I think I think this even without acausal trade.

Comment by Tamsin Leake (carado-1) on Non-Disparagement Canaries for OpenAI · 2024-05-30T19:31:37.913Z · LW · GW

sigh I wish people realized how useless it is to have money when the singularity happens. Either we die or we get a utopia in which it's pretty unlikely that pre-singularity wealth matters. What you want to maximize is not your wealth but your utility function, and you sure as hell are gonna get more from LDT handshakes with aligned superintelligences in saved worlds, if you don't help OpenAI reduce the amount of saved worlds.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-05-18T17:31:53.297Z · LW · GW

I believe that ChatGPT was not released with the expectation that it would become as popular as it did.

Well, even if that's true, causing such an outcome by accident should still count as evidence of vast irresponsibility imo.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-05-18T16:18:11.918Z · LW · GW

I'm surprised at people who seem to be updating only now about OpenAI being very irresponsible, rather than updating when they created a giant public competitive market for chatbots (which contains plenty of labs that don't care about alignment at all), thereby reducing how long everyone has to solve alignment. I still parse that move as devastating the commons in order to make a quick buck.

Comment by Tamsin Leake (carado-1) on What do you value ? · 2024-05-10T15:45:00.580Z · LW · GW

I made guesses about my values a while ago, here.

Comment by Tamsin Leake (carado-1) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:40:48.223Z · LW · GW

but that this would be bad if the users aren't one of "us"—you know, the good alignment researchers who want to use AI to take over the universe, totally unlike those evil capabilities researchers who want to use AI to produce economically valuable goods and services.

Rather, " us" — the good alignment researchers who will be careful at all about the long term effects of our actions, unlike capabilities researchers who are happy to accelerate race dynamics and increase p(doom) if they make a quick profit out of it in the short term.

Comment by Tamsin Leake (carado-1) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:30:44.183Z · LW · GW

I am a utilitarian and agree with your comment.

The intent of the post was

  • to make people weigh whether to publish or not, because I think some people are not weighing this enough
  • to give some arguments in favor of "you might be systematically overestimating the utility of publishing", because I think some people are doing that

I agree people should take the utilitalianly optimal action, I just think they're doing the utilitarian calculus wrong or not doing the calculus at all.

Comment by Tamsin Leake (carado-1) on Please stop publishing ideas/insights/research about AI · 2024-05-02T20:24:59.876Z · LW · GW

I think research that is mostly about outer alignment (what to point the AI to) rather than inner alignment (how to point the AI to it) tends to be good — quantilizers, corrigibility, QACI, decision theory, embedded agency, indirect normativity, infra bayesianism, things like that. Though I could see some of those backfiring the way RLHF did — in the hands of a very irresponsible org, even not very capabilities-related research can be used to accelerate timelines and increase race dynamics if the org doing it thinks it can get a quick buck out of it.

Comment by Tamsin Leake (carado-1) on Please stop publishing ideas/insights/research about AI · 2024-05-02T17:33:18.369Z · LW · GW

I don't buy the argument that safety researchers have unusually good ideas/research compared to capability researchers at top labs

I don't think this particularly needs to be true for my point to hold; they only need to have reasonably good ideas/research, not unusually good, for them to publish less to be a positive thing.

That said, if someone hasn't thought at all about concepts like "differentially advancing safety" or "capabilities externalities," then reading this post would probably be helpful, and I'd endorse thinking about those issues.

That's a lot of what I intend to do with this post, yes. I think a lot of people do not think about the impact of publishing very much and just blurt-out/publish things as a default action, and I would like them to think about their actions more.

Comment by Tamsin Leake (carado-1) on Please stop publishing ideas/insights/research about AI · 2024-05-02T16:08:44.709Z · LW · GW

One straightforward alternative is to just not do that; I agree it's not very satisfying but it should still be the action that's pursued if it's the one that has more utility.

I wish I had better alternatives, but I don't. But the null action is an alternative.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-05-02T07:17:31.657Z · LW · GW

It certainly is possible! In more decision-theoritic terms, I'd describe this as "it sure would suck if agents in my reference class just optimized for their own happiness; it seems like the instrumental thing for agents in my reference class to do is maximize for everyone's happiness". Which is probly correct!

But as per my post, I'd describe this position as "not intrinsically altruistic" — you're optimizing for everyone's happiness because "it sure would sure if agents in my reference class didn't do that", not because you intrinsically value that everyone be happy, regardless of reasoning about agents and reference classes and veils of ignorance.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-27T17:25:28.164Z · LW · GW

decision theory is no substitute for utility function

some people, upon learning about decision theories such as LDT and how it cooperates on problems such as the prisoner's dilemma, end up believing the following:

my utility function is about what i want for just me; but i'm altruistic (/egalitarian/cosmopolitan/pro-fairness/etc) because decision theory says i should cooperate with other agents. decision theoritic cooperation is the true name of altruism.

it's possible that this is true for some people, but in general i expect that to be a mistaken analysis of their values.

decision theory cooperates with agents relative to how much power they have, and only when it's instrumental.

in my opinion, real altruism (/egalitarianism/cosmopolitanism/fairness/etc) should be in the utility function which the decision theory is instrumental to. i actually intrinsically care about others; i don't just care about others instrumentally because it helps me somehow.

some important aspects that my utility-function-altruism differs from decision-theoritic-cooperation includes:

  • i care about people weighed by moral patienthood, decision theory only cares about agents weighed by negotiation power. if an alien superintelligence is very powerful but isn't a moral patient, then i will only cooperate with it instrumentally (for example because i care about the alien moral patients that it has been in contact with); if cooperating with it doesn't help my utility function (which, again, includes altruism towards aliens) then i won't cooperate with that alien superintelligence. corollarily, i will take actions that cause nice things to happen to people even if they've very impoverished (and thus don't have much LDT negotiation power) and it doesn't help any other aspect of my utility function than just the fact that i value that they're okay.
  • if i can switch to a better decision theory, or if fucking over some non-moral-patienty agents helps me somehow, then i'll happily do that; i don't have goal-content integrity about my decision theory. i do have goal-content integrity about my utility function: i don't want to become someone who wants moral patients to unconsentingly-die or suffer, for example.
  • there seems to be a sense in which some decision theories are better than others, because they're ultimately instrumental to one's utility function. utility functions, however, don't have an objective measure for how good they are. hence, moral anti-realism is true: there isn't a Single Correct Utility Function.

decision theory is instrumental; the utility function is where the actual intrinsic/axiomatic/terminal goals/values/preferences are stored. usually, i also interpret "morality" and "ethics" as "terminal values", since most of the stuff that those seem to care about looks like terminal values to me. for example, i will want fairness between moral patients intrinsically, not just because my decision theory says that that's instrumental to me somehow.

Comment by Tamsin Leake (carado-1) on LW Frontpage Experiments! (aka "Take the wheel, Shoggoth!") · 2024-04-26T18:52:45.280Z · LW · GW

I would feel better about this if there was something closer to (1) on which to discuss what is probably the most important topic in history (AI alignment). But noted.

Comment by Tamsin Leake (carado-1) on LW Frontpage Experiments! (aka "Take the wheel, Shoggoth!") · 2024-04-26T15:34:01.451Z · LW · GW

I'm generally not a fan of increasing the amount of illegible selection effects.

On the privacy side, can lesswrong guarantee that, if I never click on Recommended, then recombee will never see an (even anonymized) trace of what I browse on lesswrong?

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-22T20:07:17.629Z · LW · GW

Here the thing that I'm calling evil is pursuing short-term profits at the cost of non-negligeably higher risk that everyone dies.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-21T17:01:11.669Z · LW · GW

Regardless of how good their alignment plans are, the thing that makes OpenAI unambiguously evil is that they created a strongly marketed public product and, as a result, caused a lot public excitement about AI, and thus lots of other AI capabilities organizations were created that are completely dismissive of safety.

There's just no good reason to do that, except short-term greed at the cost of higher probability that everyone (including people at OpenAI) dies.

(No, "you need huge profits to solve alignment" isn't a good excuse — we had nowhere near exhausted the alignment research that can be done without huge profits.)

Comment by Tamsin Leake (carado-1) on What convincing warning shot could help prevent extinction from AI? · 2024-04-14T06:20:25.187Z · LW · GW

There's also the case of harmful warning shots: for example, if it turns out that, upon seeing an AI do a scary but impressive thing, enough people/orgs/states go "woah, AI is powerful, I should make one!" or "I guess we're doomed anyways, might as well stop thinking about safety and just enjoy making profit with AI while we're still alive", to offset the positive effect. This is totally the kind of thing that could be the case in our civilization.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-14T04:30:06.718Z · LW · GW

There could be a difference but only after a certain point in time, which you're trying to predict / plan for.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-13T20:20:10.761Z · LW · GW

What you propose, ≈"weigh indices by kolmogorov complexity" is indeed a way to go about picking indices, but "weigh indices by one over their square" feels a lot more natural to me; a lot simpler than invoking the universal prior twice.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-13T17:52:44.269Z · LW · GW

If you use the UTMs for cartesian-framed inputs/outputs, sure; but if you're running the programs as entire worlds, then you still have the issue of "where are you in time".

Say there's an infinitely growing conway's-game-of-life program, or some universal program, which contains a copy of me at infinitely many locations. How do I weigh which ones are me?

It doesn't matter that the UTM has a fixed amount of weight, there's still infinitely many locations within it.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-04-13T11:47:48.333Z · LW · GW

(cross-posted from my blog)

Is quantum phenomena anthropic evidence for BQP=BPP? Is existing evidence against many-worlds?

Suppose I live inside a simulation ran by a computer over which I have some control.

  • Scenario 1: I make the computer run the following:

    pause simulation
    
    if is even(calculate billionth digit of pi):
    	resume simulation
    

    Suppose, after running this program, that I observe that I still exist. This is some anthropic evidence for the billionth digit of pi being even.

    Thus, one can get anthropic evidence about logical facts.

  • Scenario 2: I make the computer run the following:

      pause simulation
      
      if is even(calculate billionth digit of pi):
      	resume simulation
      else:
      	resume simulation but run it a trillion times slower
    

    If you're running on the non-time-penalized solomonoff prior, then that's no evidence at all — observing existing is evidence that you're being ran, not that you're being ran fast. But if you do that, a bunch of things break including anthropic probabilities and expected utility calculations. What you want is a time-penalized (probably quadratically) prior, in which later compute-steps have less realityfluid than earlier ones — and thus, observing existing is evidence for being computed early — and thus, observing existing is some evidence that the billionth digit of pi is even.

  • Scenario 3: I make the computer run the following:

      pause simulation
    
      quantum_algorithm <- classical-compute algorithm which simulates quantum algorithms the fastest
    
      infinite loop:
      	use quantum_algorithm to compute the result of some complicated quantum phenomena
    
      	compute simulation forwards by 1 step
    

    Observing existing after running this program is evidence that BQP=BPP — that is, classical computers can efficiently run quantum algorithms: if BQP≠BPP, then my simulation should become way slower, and existing is evidence for being computed early and fast (see scenario 2).

    Except, living in a world which contains the outcome of cohering quantum phenomena (quantum computers, double-slit experiments, etc) is very similar to the scenario above! If your prior for the universe is a programs, penalized for how long they take to run on classical computation, then observing that the outcome of quantum phenomena is being computed is evidence that they can be computed efficiently.

  • Scenario 4: I make the computer run the following:

      in the simulation, give the human a device which generates a sequence of random bits
      pause simulation
    
      list_of_simulations <- [current simulation state]
    
      quantum_algorithm <- classical-compute algorithm which simulates quantum algorithms the fastest
    
      infinite loop:
      	list_of_new_simulations <- []
      	
      	for simulation in list_of_simulations:
      		list_of_new_simulations += 
      			[ simulation advanced by one step where the device generated bit 0,
      			  simulation advanced by one step where the device generated bit 1 ]
    
      	list_of_simulations <- list_of_new_simulations
    

    This is similar to what it's like to being in a many-worlds universe where there's constant forking.

    Yes, in this scenario, there is no "mutual destruction", the way there is in quantum. But with decohering everett branches, you can totally build exponentially many non-mutually-destructing timelines too! For example, you can choose to make important life decisions based on the output of the RNG, and end up with exponentially many different lives each with some (exponentially little) quantum amplitude, without any need for those to be compressible together, or to be able to mutually-destruct. That's what decohering means! "Recohering" quantum phenomena interacts destructively such that you can compute the output, but decohering* phenomena just branches.

    The amount of different simulations that need to be computed increases exponentially with simulation time.

    Observing existing after running this program is very strange. Yes, there are exponentially many me's, but all of the me's are being ran exponentially slowly; they should all not observe existing. I should not be any of them.

    This is what I mean by "existing is evidence against many-worlds" — there's gotta be something like an agent (or physics, through some real RNG or through computing whichever variables have the most impact) picking a only-polynomially-large set of decohered non-compressible-together timelines to explain continuing existing.

    Some friends tell me "but tammy, sure at step N each you has only 1/2^N quantum amplitude, but at step N there's 2^N such you's, so you still have 1 unit of realityfluid" — but my response is "I mean, I guess, sure, but regardless of that, step N occurs 2^N units of classical-compute-time in the future! That's the issue!".

Some notes:

  • I heard about pilot wave theory recently, and sure, if that's one way to get single history, why not. I hear that it "doesn't have locality", which like, okay I guess, that's plausibly worse program-complexity wise, but it's exponentially better after accounting for the time penalty.

  • What if "the world is just Inherently Quantum"? Well, my main answer here is, what the hell does that mean? It's very easy for me to imagine existing inside of a classical computation (eg conway's game of life); I have no idea what it'd mean for me to exist in "one of the exponentially many non-compressible-together decohered exponenially-small-amplitude quantum states that are all being computed forwards". Quadratically-decaying-realityfluid classical-computation makes sense, dammit.

  • What if it's still true — what if I am observing existing with exponentially little (as a function of the age of the universe) realityfluid? What if the set of real stuff is just that big?

    Well, I guess that's vaguely plausible (even though, ugh, that shouldn't be how being real works, I think), but then the tegmark 4 multiverse has to contain no hypotheses in which observers in my reference class occupy more than exponentially little realityfluid.

    Like, if there's a conway's-game-of-life simulation out there in tegmark 4, whose entire realityfluid-per-timestep is equivalent to my realityfluid-per-timestep, then they can just bruteforce-generate all human-brain-states and run into mine by chance, and I should have about as much probability of being one of those random generations as I'd have being in this universe — both have exponentially little of their universe's realityfluid! The conway's-game-of-life bruteforced-me has exponentially little realityfluid because she's getting generated exponentially late, and quantum-universe me has exponentially little realityfluid because I occupy exponentially little of the quantum amplitude, at every time-step.

    See why that's weird? As a general observer, I should exponentially favor observing being someone who lives in a world where I don't have exponentially little realityfluid, such as "person who lives only-polynomially-late into a conway's-game-of-life, but happened to get randomly very confused about thinking that they might inhabit a quantum world".

Existing inside of a many-worlds quantum universe feels like aliens pranksters-at-orthogonal-angles running the kind of simulation where the observers inside of it to be very anthropically confused once they think about anthropics hard enough. (This is not my belief.)

Comment by Tamsin Leake (carado-1) on LessWrong's (first) album: I Have Been A Good Bing · 2024-04-01T11:56:22.973Z · LW · GW

I didn't see a clear indication in the post about whether the music is AI-generated or not, and I'd like to know; was there an indication I missed?

(I care because I'll want to listen to that music less if it's AI-generated.)

Comment by Tamsin Leake (carado-1) on On expected utility, part 1: Skyscrapers and madmen · 2024-03-31T06:14:21.947Z · LW · GW

Unlike on your blog, the images on the lesswrong version of this post are now broken.

Comment by Tamsin Leake (carado-1) on Orthogonality Thesis seems wrong · 2024-03-25T13:13:59.811Z · LW · GW

Taboo the word "intelligence".

An agent can superhumanly-optimize any utility function. Even if there are objective values, a superhuman-optimizer can ignore them and superhuman-optimize paperclips instead (and then we die because it optimized for that harder than we optimized for what we want).

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-21T09:44:37.430Z · LW · GW

(I'm gonna interpret these disagree-votes as "I also don't think this is the case" rather than "I disagree with you tamsin, I think this is the case".)

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-20T13:25:50.633Z · LW · GW

I don't think this is the case, but I'm mentioning this possibility because I'm surprised I've never seen someone suggest it before:

Maybe the reason Sam Altman is taking decisions that increase p(doom) is because he's a pure negative utilitarian (and he doesn't know-about/believe-in acausal trade).

Comment by Tamsin Leake (carado-1) on Toki pona FAQ · 2024-03-18T14:40:48.072Z · LW · GW

For writing, there's also jan misali's ASCII toki pona syllabary.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-16T17:56:09.067Z · LW · GW

Reposting myself from discord, on the topic of donating 5000$ to EA causes.

if you're doing alignment research, even just a bit, then the 5000$ are probly better spent on yourself

if you have any gears level model of AI stuff then it's better value to pick which alignment org to give to yourself; charity orgs are vastly understaffed and you're essentially contributing to the "picking what to donate to" effort by thinking about it yourself

if you have no gears level model of AI then it's hard to judge which alignment orgs it's helpful to donate to (or, if giving to regranters, which regranters are good at knowing which alignment orgs to donate to)

as an example of regranters doing massive harm: openphil gave 30M$ to openai at a time where it was critically useful to them, (supposedly in order to have a chair on their board, and look how that turned out when the board tried to yeet altman)

i know of at least one person who was working in regranting and was like "you know what i'd be better off doing alignment research directly" — imo this kind of decision is probly why regranting is so understaffed

it takes technical knowledge to know what should get money, and once you have technical knowledge you realize how much your technical knowledge could help more directly so you do that, or something

Comment by Tamsin Leake (carado-1) on More people getting into AI safety should do a PhD · 2024-03-15T09:41:37.784Z · LW · GW

yes, edited

Comment by Tamsin Leake (carado-1) on More people getting into AI safety should do a PhD · 2024-03-15T05:42:02.441Z · LW · GW

So this option looks unattractive if you think transformative AI systems are likely to developed within the next 5 years. However, with a 10-years timeframe things look much stronger: you would still have around 5 years to contribute as a research.

This phrasing is tricky! If you think TAI is coming in approximately 10 years then sure, you can study for 5 years and then do research for 5 years.

But if you think TAI is coming within 10 years (for example, if you think that the current half-life on worlds surviving is 10 years; if you think 10 years is the amount of time in which half of worlds are doomed) then depending on your distribution-over-time you should absolutely not wait 5 years before doing research, because TAI could happen in 9 years but it could also happen in 1 year. If you think TAI is coming within 10 years, then (depending on your distribution) you should still in fact do research asap.

(People often get this wrong! They think that "TAI probably within X years" necessarily means "TAI in approximately X years".)

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-12T08:39:25.202Z · LW · GW

Sure, this is just me adapting the idea to the framing people often have, of "what technique can you apply to an existing AI to make it safe".

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-12T08:06:21.396Z · LW · GW

AI safety is easy. There's a simple AI safety technique that guarantees that your AI won't end the world, it's called "delete it".

AI alignment is hard.

Comment by Tamsin Leake (carado-1) on 0th Person and 1st Person Logic · 2024-03-10T05:33:04.864Z · LW · GW

I'm confused about why 1P-logic is needed. It seems to me like you could just have a variable X which tracks "which agent am I" and then you can express things like sensor_observes(X, red) or is_located_at(X, northwest). Here and Absent are merely a special case of True and False when the statement depends on X.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-09T20:59:25.892Z · LW · GW

Moral patienthood of current AI systems is basically irrelevant to the future.

If the AI is aligned then it'll make itself as moral-patient-y as we want it to be. If it's not, then it'll make itself as moral-patient-y as maximizes its unaligned goal. Neither of those depend on whether current AI are moral patients.

Comment by Tamsin Leake (carado-1) on Tamsin Leake's Shortform · 2024-03-09T16:34:29.210Z · LW · GW

If my sole terminal value is "I want to go on a rollercoaster", then an agent who is aligned to me would have the value "I want Tamsin Leake to go on a rollercoaster", not "I want to go on a rollercoaster myself". The former necessarily-has the same ordering over worlds, the latter doesn't.

Comment by Tamsin Leake (carado-1) on Let's build definitely-not-conscious AI · 2024-03-08T00:03:46.067Z · LW · GW
  • I think the term "conscious" is very overloaded and the source of endless confusion and should be tabood. I'll be answering as if the numbers are not "probability(-given-uncertainty) of conscious" but "expected(-given-uncertainty) amount of moral patienthood", calibrated with 1 meaning "as much as a human" (it could go higher — some whales have more neurons/synapses than humans and so they might plausibly be more of a moral patient than humans, in the sense that in a trolley problem you should prefer to save 1000 such whales to 1001 humans).
  • Besides the trivia I just mentioned about whales, I'm answering this mostly on intuition, without knowing off the top of my head (nor looking up) the amount of neurons/synapses. Not to imply that moral patienthood is directly linear to amount of neurons/synapses, but I expect that that amount probably matters to my notion of moral patienthood.
  • I'll also assume that everyone has a "normal amount of realityfluid" flowing through them (rather than eg being simulated slower, or being fictional, or having "double-thick neurons made of gold" in case that matters).

First list: 1, 1, 1, .7, 10⁻², 10⁻³, 10⁻⁶, 10⁻⁶, 10⁻⁸, ε, ε, ε, ε, ε.

Second list: .6, .8, .7, .7, .6, .6, .5, ε, ε, ε, ε.

Edit: Thinking about it more, something feels weird here, like these numbers don't track at all "how many of these would make me press the lever on the trolley problem vs 1 human" — for one, killing a sleeping person is about as bad as killing an awake person because like the sleeping person is a temporarily-paused-backup for an awake person. I guess I should be thinking about "the universe has budget for one more hour of (good-)experience just before heat death, but it needs to be all same species, how much do I value each?" or something.

Comment by Tamsin Leake (carado-1) on Even if we lose, we win · 2024-01-15T08:38:17.753Z · LW · GW

If you start out with CDT, then the thing you converge to is Son of CDT rather than FDT.
(that arbital page takes a huge amount of time to load for me for some reason, but it does load eventually)

And I could totally see the thing that kills us {being built with} or {happening to crystallize with} CDT rather than FDT.

We have to actually implement/align-the-AI-to the correct decision theory.

Comment by Tamsin Leake (carado-1) on Even if we lose, we win · 2024-01-15T08:32:21.455Z · LW · GW

By thinking about each other's source code, FAI and Clippy will be able to cooperate acausally like Alice and Bob, each turning their future lightcone into 10% utopia, 90% paperclips. Therefore, we get utopia either way! :D

So even if we lose we win, but even if we win we lose. The amount of utopiastuff is exactly conserved, and launching unaligned AI causes timelines-where-we-win to have less utopia by exactly as much as our timeline has more utopia.

The amount of utopiastuff we get isn't just proportional to how much we solve alignment, it's actually back to exactly equal.

See also: Decision theory does not imply that we get to have nice things.