Outcome Terminology?

post by Dach · 2020-09-14T18:04:05.048Z · LW · GW · None comments

This is a question post.

I'm writing a post about S-risks, and I need access to some clean, established terminology/background material for discussing AI-based long-term outcomes for humanity.

My current (very limited) vocabulary can be summarized with the following categories: 

  1. Outcomes which are roughly maximally bad: Hyperexistential risk/S-risk/Unfriendly AI/Existential risk
  2. Outcomes which are nontrivially worse than paperclipping-equivalents but better than approximate minimization of human utility: Hyperexistential risk/S-risk/Unfriendly AI/Existential risk
  3. Outcomes which are produced by agents essentially orthogonal to human values: Paperclipping/Unfriendly AI/Existential risk
  4. Outcomes which are nontrivially better than paperclipping but worse than Friendly AI: ???
  5. Outcomes which are roughly maximally good: Friendly AI

The problems are manifold: 

I've done a small amount of investigation and determined less brainpower would be wasted by just asking for links.


None comments

Comments sorted by top scores.