Posts

Comments

Comment by Tarnish on Is there a known method to find others who came across the same potential infohazard without spoiling it to the public? · 2024-10-17T11:29:10.839Z · LW · GW

As far as I know, there is unfortunately no system for this. I think what people typically do is contact MIRI leadership, but I don't know MIRI leadership to have particularly put silent people in touch with other silent people as a result.

Comment by Tarnish on Most arguments for AI Doom are either bad or weak · 2024-10-12T22:58:05.179Z · LW · GW

Strong arguments of this kind? I sure hope not, that'd make it easier for more people to find insights for how to build an AI that causes doom.

Comment by Tarnish on Most arguments for AI Doom are either bad or weak · 2024-10-12T12:36:17.779Z · LW · GW

Note that some of the best arguments are of the shape "AI will cause doom because it's not that hard to build the following..." followed by insights about how to build an AI that causes doom. Those arguments are best rederived privately rather than shared publicly, and by asking publicly you're filtering the strength of arguments you might get exposed to.

Comment by Tarnish on Provably Safe AI: Worldview and Projects · 2024-08-10T04:44:00.293Z · LW · GW

Unfortunately, that does not appear to be a stable solution. Even if the US paused its AI development, China or other countries could gain an advantage by accelerating their own work.

Arguing-for-pausing does not need to be a stable solution to help. If it buys time, that's already helpful. If the US pauses AI development, but China doesn't, that's still less many people working on AI that might kill everyone.

Comment by Tarnish on Petrov Day [Spoiler Warning] · 2023-09-27T21:29:40.127Z · LW · GW

Mu. The most basic rationalist precept is to not forcibly impose your values onto another mind.

It is? Last I heard, the two most basic precepts of rationality were:

  1. Epistemic rationality: systematically improving the accuracy of your beliefs.
  2. Instrumental rationality: systematically achieving your values.

(Typically with a note saying "ultimately, when at odds, the latter trumps the former")