Why do you reject negative utilitarianism?
post by Teo Ajantaival · 2019-02-11T15:38:39.163Z · LW · GW · 9 commentsThis is a question post.
Contents
Answers 48 Rob Bensinger 35 Kaj_Sotala 28 Wei_Dai 27 Lukas_Gloor 16 Teo Ajantaival 11 ESRogs 8 Andaro 4 xdxdxd None 9 comments
Note (2022-03-15): For a recent defense of similar axiological views, see my series on minimalist axiologies [? · GW].
(Crossposted on the EA Forum [EA · GW])
Absolute negative utilitarianism (ANU) is a minority view despite the theoretical advantages of terminal value monism (suffering is the only thing that motivates us “by itself”) over pluralism (there are many such things). Notably, ANU doesn’t require solving value incommensurability, because all other values can be instrumentally evaluated by their relationship to the suffering of sentient beings, using only one terminal value-grounded common currency for everything.
Therefore, it is a straw man argument that NUs don’t value life or positive states, because NUs value them instrumentally, which may translate into substantial practical efforts to protect them (compared even with someone who claims to be terminally motivated by them).
If the rationality and EA communities are looking for a unified theory of value, why are they not converging (more) on negative utilitarianism?
What have you read about it that has caused you to stop considering it, or to overlook it from the start?
Can you teach me how to see positive states as terminally (and not just instrumentally) valuable, if I currently don’t? (I still enjoy things, being closer to the extreme of hyperthymia than anhedonia. Am I platonically blind to the intrinsic aspect of positivity?)
And if someone wants to answer: What is the most extreme form of suffering that you’ve experienced and believe can be “outweighed” by positive experiences?
Answers
I find negative utilitarianism unappealing for roughly the same reason I'd find "we should only care about disgust" or "we should only care about the taste of bananas" unappealing. Or if you think suffering is much closer to a natural kind than disgust, then supply some other mental (or physical!) state that seems more natural-kind-ish to you.
"Only suffering ultimately matters" and "only the taste of bananas ultimately matters" share the virtue of simplicity, but they otherwise run into the same difficulty, which is just that they don't exhaustively describe all the things I enjoy or want or prefer. I don't think my rejection of bananatarianism has to be any more complicated than that.
Something I wrote last year in response to a tangentially related paper:
I personally care about things other than suffering. What are negative utilitarians saying about that?
Are they saying that they don't care about things like friendship, good food, joy, catharsis, adventure, learning new things, falling in love, etc., except as mechanisms for avoiding suffering? Are they saying that I'm deluded about having preferences like those? Are they saying that I should try to change my preferences — and if so, why? Are they saying that my preferences are fine in my personal decision-making as an individual, but shouldn't get any weight in an idealized negotiation about what humanity as a group should do (ignoring any weight my preferences get from non-NU views that might in fact warrant a place at the bargaining table for more foundational or practical reasons distinct from the NU ideal) — and if so, why?
[...] "It's wrong to ever base any decision whatsoever on my (or anyone else's) enjoyment of anything whatsoever in life, except insofar as that enjoyment has downstream effects on other things" is an incredibly, amazingly strong claim. And it's important in this context that you're actually making that incredibly strong claim: more mild "negative-leaning" utilitarianisms (which probably shouldn't be associated with NU, given how stark the difference is) don't have to deal with the version of the world destruction argument I think x-risk people tend to be concerned about, which is not 'in some scenarios, careful weighing of the costs and benefits can justify killing lots of people' but rather 'any offsets or alternatives to building misaligned resource-hungry AGI (without suffering subsystems) get literally zero weight, if you're sufficiently confident that that's what you're building; there's no need to even consider them; they aren't even a feather on the scale'. I just don't see why the not-even-a-feather-on-the-scale view deserves any more attention or respect than, e.g., divine-command theory; in an argument between the "negative-leaning" utilitarian and the real negative utilitarian, I don't think the NU gets any good hits in.
(Simplicity is a virtue, but not when it's of the "I'm going to attempt to disregard every consideration in all of my actions going forward except the expected amount of deliciousness in the future" or "... except the expected amount of lying in the future" variety; so simplicity on its own doesn't raise the view to the level of having non-negligible probability compared to negative-learning U.)
↑ comment by Teo Ajantaival · 2019-02-12T20:28:33.299Z · LW(p) · GW(p)
Yes, I am making the (AFAICT, in your perspective) “incredibly, amazingly strong claim” that in a unified theory, only suffering ultimately matters. In other words, impartial compassion is the ultimate scale (comparator) to decide conflicts between expected suffering vs. other values (whose common basis for this comparison derives from their complete, often context-dependent relationship to expected suffering, including accounting for the wider incentives & long-term consequences from breaking rules that are practically always honored).
I find negative utilitarianism unappealing for roughly the same reason I'd find "we should only care about disgust" or "we should only care about the taste of bananas" unappealing.
Roughly? Suffering is not an arbitrary foundation for unification (for a “common currency” underlying an ideally shared language for cause prioritization). Suffering is the clearest candidate for a thing we all find terminally motivating, at least once we know what we’re talking about (i.e., aren’t completely out of touch with the nature of extreme suffering, as evaluators & comparators of experiences are expected not to be). Personally, I avoid the semantics of arguing over what we “should” care about. Instead, I attempt to find out what I do care about, what these values’ motivational currency is ultimately derived from, and how could I unify these findings into a psychologically realistic model with consistent practical implications & minimal irreconcilable contradictions (such as outweighing between multiple terminal values, because I’m a skeptic of aggregation over space and time; aggregate experiences physically exist only as thoughts not fit to outweigh suffering, which only preventing more suffering can do).
"Only suffering ultimately matters" and "only the taste of bananas ultimately matters" share the virtue of simplicity, but they otherwise run into the same difficulty, which is just that they don't exhaustively describe all the things I enjoy or want or prefer. I don't think my rejection of bananatarianism has to be any more complicated than that.
I agree re: bananatarianism, but there’s more to unpack from the suffering-motivated unification than meets the eye.
No verbal descriptions can exhaustively describe all the things we enjoy, want, or prefer, because our inner homeostatic & psychological dynamics contain processes that are too multidimensional for simple overarching statements. What we can do, is unpack the implications of {“Only suffering ultimately matters”} to see how this can imply, predict, retrodict, and explain our other motivations.
In evolutionary and developmental history terms, we can see at the first quick glance that many (if not immediately all) of our other motivations interact with suffering, or have interacted with our suffering in the past (individually, neurally, culturally, evolutionarily). They serve functions of group cohesion, coping with stress, acquiring resources, intimacy, adaptive learning & growth, social deterrence, self-protection, understanding ourselves, and various other things we value & honor because they make life easier or interesting. Neither NU nor other systems will honor all of our perceived wants as absolutes to maximize (reproduction) or to even respect at all (revenge; animalistic violence; desire for personal slaves or worship), but most of our intuitively nominated “terminal” values need not be overriden by the slightest suffering, because they do serve strong functions to prevent suffering, especially when they seem to us like autonomous goals without constantly reminding us of how horrible things did and would happen without them. NU simply claims that it is the most diplomatic solution for a unified theory to de-attach from other values as absolutes, and respect them to the degree that we need them (practicing epistemic uncertainty when we do not yet understand the full role of something we intuitively deeply value!). This may in practice lead to great changes, ideally in directions of more self-compassion and general compassion for others without our other values overriding our motivation to prevent as many cases of extreme suffering as we can.
A considerate rejection of NU needs to be more complicated than of bananatarianism, because the unity, applicability, and explanatory power of NU relies on its implications (instead of explicit absolute rules or independent utility assignments for exemplary grounding-units of every value—challenges for other systems), and its weights of instrumental value depend not on static snapshots of worlds, but on the full context and counterfactuals epistemically accessible to us in each situation. In extreme situations, we may decide it worthwhile to simulate the expected long-term consequences of possibly bending rules that we normally accept as near-absolute heuristics to save ourselves from resource-intensive overthinking (e.g., the degree to which we respect someone’s autonomy, in emergencies). This doesn’t imply that superweapon research is a low-hanging fruit for aspiring suffering-minimizers (for reasons I won’t detail right now, because I find the world-wiping objections worth addressing mostly in the context of AGI assumptions; my primary interest here is unification for practical cause prioritization, worth noting).
To actually reject NU, you must explain what makes something (other than suffering) terminally valuable (or as I say, motivating) beyond its instrumental value for helping us prevent suffering in the total context. This instrumental value is multifaceted and can be derived from various kinds of relationships to suffering. So other “terminal” values may serve important purposes, including that they help us (some examples in parentheses):
- cope with suffering (coping mechanisms, friendship, community)
- avoid ruminating on suffering (explicit focus on expansive, positive language and goals that don’t contain reminders of their possibly suffering-mediated usefulness)
- re-interpret suffering (humor, narratives, catharsis)
- prevent suffering (science, technology, cognitive skills & tools)
- understand suffering (wide & deep personal experience, culture)
- predict suffering (science)
- skip epistemic difficulties of trying to optimize others’ suffering for them (autonomy)
- prevent the abuse of our being motivated by suffering (human rights, justice system, deterrence)
- help others’ suffering (life, health, freedom, personal extra resources to invest as we see fit, reducing x-risk)
- resilience against suffering (experience, intelligence, learning, cultural memory)
- safety against emergencies (family, intimacy, community, institutions)
- help us relax and regain our ability to help (good food, joy, replenishing activities)
- avoid a horrible, anxiety-epidemic spreading societal collapse (not getting caught secretly killing people and everyone who knew them, in the name of compassion, by not doing this)
To reject NU, is there some value you want to maximize beyond self-compassion and its role for preventing suffering, at the risk of allowing extreme suffering? How will you tell this to someone undergoing extreme suffering?
NUs are not saying you are deluded for valuing multiple things. But you may be overly attached to them if you—beyond self-compassion—would want to spend your attention on copying/boosting instances of them rather than on preventing others from having to undergo extreme suffering.
After writing this, I wonder if the actual disagreement is still the fear that an NU-AGI would consider humans less {instrumentally valuable for preventing suffering} than it would consider {our suffering terminally worth preventing}. This feels like a very different conversation than what would be a useful basis for a common language of cause prioritization.
Replies from: RobbBB, Raemon↑ comment by Rob Bensinger (RobbBB) · 2019-02-13T00:42:21.197Z · LW(p) · GW(p)
In evolutionary and developmental history terms, we can see at the first quick glance that many (if not immediately all) of our other motivations interact with suffering, or have interacted with our suffering in the past (individually, neurally, culturally, evolutionarily). They serve functions of group cohesion, coping with stress, acquiring resources, intimacy, adaptive learning & growth, social deterrence, self-protection, understanding ourselves, and various other things we value & honor because they make life easier or interesting.
Seems like all of this could also be said of things like "preferences", "enjoyment", "satisfaction", "feelings of correctness", "attention", "awareness", "imagination", "social modeling", "surprise", "planning", "coordination", "memory", "variety", "novelty", and many other things.
"Preferences" in particular seems like an obvious candidate for 'thing to reduce morality to'; what's your argument for only basing our decisions on dispreference or displeasure and ignoring positive preferences or pleasure (except instrumentally)?
Neither NU nor other systems will honor all of our perceived wants as absolutes to maximize
I'm not sure I understand your argument here. Yes, values are complicated and can conflict with each other. But I'd rather try to find reasonable-though-imperfect approximations and tradeoffs, rather than pick a utility function I know doesn't match human values and optimize it instead just because it's uncomplicated and lets us off the hook for thinking about tradeoffs between things we ultimately care about.
E.g., I like pizza. You could say that it's hard to list every possible flavor I enjoy in perfect detail and completeness, but I'm not thereby tempted to stop eating pizza, or to try to reduce my pizza desire to some other goal like 'existential risk minimization' or 'suffering minimization'. Pizza is just one of the things I like.
To actually reject NU, you must explain what makes something (other than suffering) terminally valuable (or as I say, motivating) beyond its instrumental value for helping us prevent suffering in the total context
E.g.: I enjoy it. If my friends have more fun watching action movies than rom-coms, then I'll happily say that that's sufficient reason for them to watch more action movies, all on its own.
Enjoying action movies is less important than preventing someone from being tortured, and if someone talks too much about trivial sources of fun in the context of immense suffering, then it makes sense to worry that they're a bad person (or not sufficiently in touch with their compassion).
But I understand your position to be not "torture matters more than action movies", but "action movies would ideally have zero impact on our decision-making, except insofar as it bears on suffering". I gather that from your perspective, this is just taking compassion to its logical conclusion; assigning some more value to saving horrifically suffering people than to enjoying a movie is compassionate, so assigning infinitely more value to the one than the other seems like it's just dialing compassion up to 11.
One reason I find this uncompelling is that I don't think the right way to do compassion is to ignore most of the things people care about. I think that helping people requires doing the hard work of figuring out everything they value, and helping them get all those things. That might reduce to "just help them suffer less" in nearly all real-world decisions nowadays, because there's an awful lot of suffering today; but that's a contingent strategy based on various organisms' makeup and environment in 2019, not the final word on everything that's worth doing in a life.
To reject NU, is there some value you want to maximize beyond self-compassion and its role for preventing suffering, at the risk of allowing extreme suffering? How will you tell this to someone undergoing extreme suffering?
I'll tell them I care a great deal about suffering, but I don't assign literally zero importance to everything else.
NU people I've talked to often worry about scenarios like torture vs. dust specks, and that if we don't treat happiness as literally of zero value, then we might make the wrong tradeoff and cause immense harm.
The flip side is dilemmas like:
Suppose you have a chance to push a button that will annihilate all life in the universe forever. You know for a fact that if you don't push it, then billions of people will experience billions upon billions of years of happy, fulfilling, suffering-free life, filled with richness, beauty, variety, and complexity; filled with the things that make life most worth living, and with relationships and life-projects that people find deeply meaningful and satisfying.
However, you also know for a fact that if you don't push the button, you'll experience a tiny, almost-unnoticeable itch on your left shoulder blade a few seconds later, which will be mildly unpleasant for a second or two before the Utopian Future begins. With this one exception, no suffering will ever again occur in the universe, regardless of whether you push the button. Do you push the button, because your momentary itch matters more than all of the potential life and happiness you'd be cutting out?
Replies from: RobbBB↑ comment by Rob Bensinger (RobbBB) · 2019-02-13T00:45:16.259Z · LW(p) · GW(p)
And if you say "I don't push the button, but only because I want to cooperate with other moral theorists" or "I don't push the button, but only because NU is very very likely true but I have nonzero moral uncertainty": do you really think that's the reason? Does that really sound like the prescription of the correct normative theory (modulo your own cognitive limitations and resultant moral uncertainty)? If the negotiation-between-moral-theories spat out a slightly different answer, would this actually be a good idea?
↑ comment by Raemon · 2019-02-12T21:34:57.980Z · LW(p) · GW(p)
This comment doesn't seem to sufficiently engage with (what I saw as) the core question Rob was asking (and which I would ask), which was:
I personally care about things other than suffering. What are negative utilitarians saying about that?
Are they saying that they don't care about things like friendship, good food, joy, catharsis, adventure, learning new things, falling in love, etc., except as mechanisms for avoiding suffering? Are they saying that I'm deluded about having preferences like those? Are they saying that I should try to change my preferences — and if so, why?
You briefly note "you may be overly attached to them", but this doesn't give any arguments for why I might be overly attached to them, instead of attached to them the correct amount.
When you ask:
To actually reject NU, you must explain what makes something (other than suffering) terminally valuable (or as I say, motivating) beyond its instrumental value for helping us prevent suffering in the total context.
My response is "to reject NU, all I have to do is terminally care about anything other than suffering. I care about things other than suffering, ergo NU must be false, and the burden is on other people to explain what is wrong with my preferences."
I used to consider myself NU, but have since then rejected it.
Part of my rejection was that, on a psychological level, it simply didn't work for me. The notion that everything only has value to the extent that it reduces suffering meant that most of the things which I cared about, were pointless and meaningless except for their instrumental value in reducing my suffering or making me more effective at reducing suffering. Doing things which I enjoyed, but constantly having a nagging sensation of "if I could just learn to no longer need this, then it would be better for everyone" basically meant that it was very hard to ever enjoy anything. It was basically setting my mind up to be a battlefield, dominated by an NU faction trying to suppress any desires which did not directly contribute to reducing suffering, and opposed by an anti-NU faction which couldn't do much but could at least prevent me from getting any effective NU work done, either.
Eventually it became obvious that even from an NU perspective, it would be better for me to stop endorsing NU, since that way I might end up actually accomplishing more suffering reduction than if I continued to endorse NU. And I think that this decision was basically correct.
A related reason is that I also rejected the need for a unified theory of value. I still think that if you wanted to reduce human values into a unified framework, then something like NU would be one of the simplest and least paradoxical answers. But eventually I concluded that any simple unified theory of value is likely to be wrong, and also not particularly useful for guiding practical decision-making. I've written more about this here [LW · GW].
Finally, and as a more recent development, I notice that NU neglects to take into account non-suffering-based preferences. My current model of minds and suffering is that minds are composed of many different subagents [? · GW] with differing goals; suffering is the result of the result of different subagents being in conflict (e.g. if one subagent wants to push through a particular global belief update, which another subagent does not wish to accept).
This means that I could imagine an advanced version of myself who had gotten rid of all personal suffering, but was still motivated by pursue other goals. Suppose for the sake of argument that I only had subagents which cared about 1) seeing friends 2) making art. Now if my subagents reached agreement of spending 30% of their time making art and 70% of their time seeing friends, then this could in principle eliminate my suffering by removing subagent conflict, but it would still be driving me to do things for reasons other than reducing suffering. Thus the argument that suffering is the only source of value fails; the version of me which had eliminated all personal suffering might be more driven to do things than the current one! (since subagent conflict was no longer blocking action in any situation)
As a practical matter, I still think that reducing suffering is one of the most urgent EA priorities: as long as death and extreme suffering exist in the world, anything that would be called "altruism" should focus its efforts on reducing that. But this is a form of prioritarianism, not NU. I do not endorse NU's prescription that an entirely dead world would be equally good or better as a world with lots of happy entities, simply because there are subagents within me who would prefer to exist and continue to do stuff, and also for other people to continue to exist and do stuff if they so prefer. I want us to liberate people's minds from involuntary suffering, and then to let people do whatever they still want to do when suffering is a thing that people experience only voluntarily.
↑ comment by MichaelStJules · 2019-10-18T17:52:49.091Z · LW(p) · GW(p)
I think most of this is compatible with preference utilitarianism (or consequentialism generally), which, in my view, is naturally negative. Nonnegative preference utilitarianism would hold that it could be good to induce preferences in others just to satisfy these preferences, which seems pretty silly.
↑ comment by Teo Ajantaival · 2019-02-13T14:47:45.686Z · LW(p) · GW(p)
Thanks for the perspective.
I agree that even NU may imply rejecting NU in its present form, because it does not feel like a psychologically realistic theory to constantly apply in everyday life; we are more motivated to move towards goals and subgoals that do not carry explicit reminders of extreme suffering on the flip side.
I do feel that I am very close to NU whenever I consider theoretical problems and edge-cases that would outweigh extreme suffering with anything else than preventing more extreme suffering. In practice, it may be more applicable (and ironically, more useful according to NU) to replace NU with asymptotically approaching uncompromising compassion, or an equivalent positive-language narrative, while taking huge grains of epistemic humility whenever it starts to feel like a good idea to override any explicit preferences of others.
Epistemic humility is also why I cannot imagine a situation where I would push The Button, because I cannot imagine how I could know with a certainty that it works (so I cannot derive an intuition out of that thought experiment).
I believe the most often cited (in the LW/EA communities) paper arguing against NU is Toby Ord's Why I'm Not a Negative Utilitarian. This and this seem to be the main replies to it from NU perspectives. (I think I've skimmed some of these articles but have not actually considered the arguments carefully.)
Ethical theories don't need to be simple. I used to have the belief that ethical theories ought to be simple/elegant/non-arbitrary for us to have a shot at them being the correct theory, a theory that intelligent civilizations with different evolutionary histories would all converge on. This made me think that NU might be that correct theory. Now I’m confident that this sort of thinking was confused: I think there is no reason to expect that intelligent civilizations with different evolutionary histories would converge on the same values, or that there is one correct set of ethics that they "should" converge on if they were approaching the matter "correctly". So, looking back, my older intuition feels confused now in a similar way as ordering the simplest food in a restaurant in expectation of anticipating what others would order if they also thought that the goal was that everyone orders the same thing. Now I just want to order the "food" that satisfies my personal criteria (and these criteria do happen to include placing value on non-arbitrariness/simplicity/elegance, but I’m a bit less single-minded about it).
Your way of unifying psychological motivations down to suffering reduction is an "externalist" account of why decisions are made, which is different from the internal story people tell themselves. Why think all people who tell different stories are mistaken about their own reasons? The point "it is a straw man argument that NUs don’t value life or positive states“ is unconvincing, as others have already pointed out. I actually share your view that a lot of things people do might in some way trace back to a motivating quality in feelings of dissatisfaction, but (1) there are exceptions to that (e.g., sometimes I do things on auto-pilot and not out of an internal sense of urgency/need, and sometimes I feel agenty and do things in the world to achieve my reflected life goals rather than tend to my own momentary well-being), and (2) that doesn’t mean that whichever parts of our minds we most identify with need to accept suffering reduction as the ultimate justification of their actions. For instance, let’s say you could prove that a true proximate cause why a person refused to enter Nozick’s experience machine was that, when they contemplated the decision, they felt really bad about the prospect of learning that their own life goals are shallower and more self-centered than they would have thought, and *therefore* they refuse the offer. Your account would say: "They made this choice driven by the avoidance of bad feelings, which just shows that ultimately they should accept the offer, or choose whichever offer reduces more suffering all-things-considered.“ Okay yeah, that's one story to tell. But the person in question tells herself the story that she made this choice because she has strong aspirations about what type of person she wants to be. Why would your externally-imported justification be more valid (for this person's life) than her own internal justification?
Thanks for the replies, everyone!
I don’t have the time to reply back individually, but I read them all and believe these to be pretty representative of the wider community’s reasons to reject NU as well.
I can’t speak for those who identify strictly as NU, but while I currently share many of NU’s answers to theoretical outweighing scenarios, I do find it difficult to unpack all the nuance it would take to reconcile “NU as CEV” with our everyday experience.
Therefore, I’ll likely update further away from
{attempting to salvage NU’s reputation by bridging it with compassion, motivation theory, and secular Buddhism}
towards
{integrating these independent of NU, seeing if this would result in a more relatable language, or if my preferred kind of theoretical unity (without pluralist outweighing) would still have the cost of its sounding absurd and extreme on its face}
↑ comment by ESRogs · 2019-02-14T19:36:58.335Z · LW(p) · GW(p)
Did you make any update regarding the simplicity / complexity of value?
My impression is that theoretical simplicity is a major driver of your preference for NU, and also that if others (such as myself) weighed theoretical simplicity more highly that they would likely be more inclined towards NU.
In other words, I think theoretical simplicity may be a double crux [LW · GW] in the disagreements here about NU. Would you agree with that?
Replies from: Teo Ajantaival↑ comment by Teo Ajantaival · 2019-02-15T12:03:16.846Z · LW(p) · GW(p)
Did you make any update regarding the simplicity / complexity of value?
Yes, in terms of how others may explicitly defend the terminal value of even preferences (tastes, hobbies), instead of defending only terminal virtues (health, friendship), or core building blocks of experience (pleasure, beauty).
No, in terms of assigning anything {independent positive value}.
I experience all of the things quoted in Complexity of value,
"Life, consciousness, and activity; health and strength; pleasures and satisfactions of all or certain kinds; happiness, beatitude, contentment, etc.; truth; knowledge and true opinions of various kinds, understanding, wisdom; beauty, harmony, proportion in objects contemplated; aesthetic experience; morally good dispositions or virtues; mutual affection, love, friendship, cooperation; just distribution of goods and evils; harmony and proportion in one's own life; power and experiences of achievement; self-expression; freedom; peace, security; adventure and novelty; and good reputation, honor, esteem, etc."
but I don’t know how to ultimately prioritize between them unless they are commensurable. I make them commensurable by weighing their interdependent value in terms of the one thing we all(?) agree is an independent motivation: preventable suffering. (If preventable suffering is not worth preventing for its own sake, what is it worth preventing for, and is this other thing agreeable to someone undergoing the suffering as the reason for its motivating power?) This does not mean that I constantly think of them in these terms (that would be counterproductive), but in conflict resolution I do not assign them independent positive numerical values, which pluralism would imply one way or another.
Any pluralist theory begs the question of outweighing suffering with enough of any independently positive value. If you think about it for five minutes, aggregate happiness (or any other experience) does not exist. If our first priority is to prevent preventable suffering, that alone is an infinite game; it does not help to make a detour to boost/copy positive states unless this is causally connected to preventing suffering. (Aggregates of suffering do not exist either, but each moment of suffering is terminally worth preventing, and we have limited attention, so aggregates and chain-reactions of suffering are useful tools of thought for preventing as many as we can. So are many other things without requiring our attaching them independent positive value, or else we would be tiling Mars with them whenever it outweighed helping suffering on Earth according to some formula.)
My experience so far with this kind of unification is that it avoids many (or even all) of the theoretical problems that are still considered canonical challenges for pluralist utilitarianisms that assign both independent negative value to suffering and independent positive value to other things. I do not claim that this would be simple or intuitive – that would be analogous to reading about some Buddhist system, realizing its theoretical unity, and teleporting past its lifelong experiential integration – but I do claim that a unified theory with grounding in a universally accepted terminal value might be worth exploring further, because we cannot presuppose that any kind of CEV would be intuitive or easy to align oneself with.
My impression is that theoretical simplicity is a major driver of your preference for NU, and also that if others (such as myself) weighed theoretical simplicity more highly that they would likely be more inclined towards NU.
In other words, I think theoretical simplicity may be a double crux [LW · GW] in the disagreements here about NU. Would you agree with that?
Partly, yes. It may also be that all of us, me included, are out of touch with the extreme ends of experience and thus do not understand the ability of some motivations to override everything else.
It is also difficult to operationalize a false belief in independent value: When are we attached to a value to the extent that we would regret not spending its resources elsewhere, on CEV-level reflection?
People also differ along their background assumptions on whether AGI makes the universally life-preventing button a relevant question, because for many, the idea of an AGI represents an omnipotent optimizer that will decide everything about the future. If so, we want to be careful about assigning independent positive value to all the things, because each one of those invites this AGI to consider {outweighing suffering} with {producing those things}, since pluralist theories do not require a causal connection between the things being weighed.
If the rationality and EA communities are looking for a unified theory of value
Are they? Many of us seem to have accepted that our values are complex.
Absolute negative utilitarianism (ANU) is a minority view despite the theoretical advantages of terminal value monism (suffering is the only thing that motivates us “by itself”) over pluralism (there are many such things). Notably, ANU doesn’t require solving value incommensurability, because all other values can be instrumentally evaluated by their relationship to the suffering of sentient beings, using only one terminal value-grounded common currency for everything.
This seems like an argument that it would be convenient if our values were simple. This does not seem like strong evidence that they actually are simple. (Though I grant that you could make an argument that it might be better to try to achieve only part of what we value if we're much more likely to be successful that way.)
What have you read about it that has caused you to stop considering it, or to overlook it from the start?
I reject impartiality on the grounds that I'm a personal identity and therefore not impartial. The utility of others is not my utility, therefore I am not a utilitarian. I reject unconditional altruism in general for this reason. It amazes me in hindsight that I was ever dumb enough to think otherwise.
Can you teach me how to see positive states as terminally (and not just instrumentally) valuable, if I currently don’t?
Teach, no, but there are some intuitions that can be evoked. I'd personally take a 10:1 ratio between pleasure and pain; if I get 10 times more pleasure out of something, I'll take any pain as a cost. It's just usually not realistic, which is why I don't agree that life has generally positive value.
There are fictional descriptions of extreme pleasure enhancement and wireheading, e.g. in fantasy that describe worthwhile states of experience. The EA movement is fighting against wireheading, as you can see in avturchin's posts. But I think such a combination of enhancement + wireheading could plausibly come closest to delivering net-positive value of life, if it could be invented (although I don't expect it in my lifetime, so it's only theoretical). Here's an example from fiction:
"You see, I have a very special spell called the Glow. It looks like this." The mage flicked his fingers and they started Glowing in a warm, radiant light. Little Joanna looked at them in awe. "It's so pretty! Can you teach me that? Is that the reward?" Melchias laughed. "No. The true reward happens when I touch you with it." She stared at him curiously. He looked down at the table in front of him with an expectant look, and she put her slender arm there so he could touch her. "Here, let me demonstrate. Now, this won't hurt one bit..." He reached out with his Glowing fingers to touch the back of her small hand, ever so gently.
And as their skin connected, Joanna's entire world exploded. The experience was indescribable. Nothing, no words and no warnings, could have prepared Joanna for the feeling that was now blasting through her young mind with the screaming ferocity of a white-hot firestorm, ripping her conscious thoughts apart like a small, flickering candlelight in a gigantic hurricane of whipping flames and shredding embers. She had no idea, no idea such pleasure existed, ever could exist! It was all of her happy memories, all of her laughter, her playfulness, her sexy tingles when she rubbed herself between the legs, the goodness of eating apple pie, the warmth of the fireplace in the winter nights, the love in her papa's strong arms, the fun of her games and friendships with the other village kids, the excitement of stories yet unheard, the exhilaration of running under the summer sun, the fascination of the nature and the animals around her, the smells and the tastes, the hugs and awkward kisses, all the goodness of all her young life, all condensed into a mere thousandth of a split-second... ...and amplified a thousand-fold... ...and shot through her mind, through her soul, again and again and again, split-second after split-second after split-second, like a bombardment of one supernova after another supernova of pure, unimaginable bliss, again and again and again, and yet again, second after second, filling her up, ripping her apart with raw ecstatic brilliance, melting her mind together in a new form, widened and brighter than it had ever been, a new, divine, god-like Joanna that no words could adequately worship, only to rip her apart again with a new fiery pulse of condensed, sizzling-hot vibrance, indescribable, unimaginable, each second an unreached peak, a new high, a new universe of fantastic pleasure, a new, unspeakably wonderful Joanna, loved and pulsing in her own Glowing light with a beauty unmatched by any other thing in all of the World Tree. She was a giant beating heart that was also a Goddess, Glowing and pulsing in the center of Everything, Glowing with the certainty of absolute affirmation, the purity of absolute perfection, the undeniability of absolute goodness. She spent centuries in seconds, serene yet in screaming ecstasy, non-living yet pulsing raw life force, non-personal yet always Joanna, Joanna in divine totality. It took Joanna a long time to realize she was still breathing, a living child with an actual human body. She had forgotton to breathe, and was sure she would have suffocated by now, but somehow, inexplicably, her body had remembered to keep itself alive without her. Master Melchias had lied: It did hurt, her chin and lip hurt, but the young girl found it was only because she had dropped to the hard stone floor in helpless twitching convulsions, and she had accidentally bitten herself. As promised by the wizard, the small wound was quickly healing. Joanna couldn't get up yet. She had no idea how much time had passed, but she just couldn't move or even open her young eyes yet. She curled up into a fetal position on the cold, hard floor of Melchias' Tower and sobbed uncontrollably. She sobbed and cried, and sobbed, and laughed madly, then sobbed and cried again. They were tears of pure joy.
The Glow wasn't just a normal pleasure spike, like an orgasm, a fit of laughter or a drug high. It went far, far beyond that. Normal human experiences existed within an intensity range that was given by nature. It served to motivate the organism for survival and reproduction, but it was not optimized for the experience itself. Even the most intense experiences, like burning alive or being skinned alive, existed within that ordinary, natural range. But the magic of the Glow didn't just stimulate pleasure within that range - it completely changed the range itself. It broke the scale on which normal experiences were measured, and then attached a vast multitude of additional ones to its top. By enhancing the part of the subject's mind that contained ordinary pleasure, it became temporarily able to experience an intensity that was hundreds of thousands times stronger than even the most extreme natural human feeling. Being drowned in hot oil, being flayed alive or tortured with needles, deep romantic love and fulfillment, orgasmic ecstasy, perfect fits of laughter - all of these human extremes represented only a miniscule fraction of the new potential. And only then did the spell induce raw, optimized pleasure within this new, widened consciousness. The result was an unimaginably pure goodness that fell so far outside of the subject's prior experience that it couldn't even be communicated by words. It had to be demonstrated. Once a potential [...] candidate had perceived even one second of the Glow, each containing more joy and happiness than an average human lifespan, with none of its pain, they all became devoted followers to Melchias. He transformed their experience from something human to something divine, and in turn, he became like a god to them.
↑ comment by Rob Bensinger (RobbBB) · 2019-02-13T21:41:04.872Z · LW(p) · GW(p)
The utility of others is not my utility, therefore I am not a utilitarian. I reject unconditional altruism in general for this reason.
When I say that I'm a utilitarian (or something utilitarian-ish), I mean something like: If there were no non-obvious bad side-effects — e.g., it doesn't damage my ability to have ordinary human relationships in a way that ends up burning more value than it creates — I'd take a pill that would bind my future self to be unwilling to sacrifice two strangers to save a friend (or to save myself), all else being equal.
The not-obviously-confused-or-silly version of utilitarianism is "not reflectively endorsing extreme partiality toward yourself or your friends relative to strangers," rather than "I literally have no goals or preferences or affection for anything other than perfectly unbiased maximization of everyone's welfare'".
If you flip the Rachels-Temkin spectrum argument (philpapers.org/archive/NEBTGT.pdf), then some tradeoff between happiness and suffering is needed to keep transitive preferences, which is necessary to avoid weird conclusions like accepting suffering to avoid happiness. As long as you don't think theres some suffering threshold where 1 more util of suffering is infinitely worse than anything else, then this makes sense.
Also NU in general has a bad reputation in the philosophy community (more than classical utilitarianism I think) so it's better EAs don't endorse it.
↑ comment by Teo Ajantaival · 2019-02-12T16:35:50.108Z · LW(p) · GW(p)
If you flip the Rachels-Temkin spectrum argument (philpapers.org/archive/NEBTGT.pdf), then some tradeoff between happiness and suffering is needed to keep transitive preferences, which is necessary to avoid weird conclusions like accepting suffering to avoid happiness. As long as you don't think theres some suffering threshold where 1 more util of suffering is infinitely worse than anything else, then this makes sense.
Can you give a practical example of a situation where I would be hereby forced to admit that happiness has terminal value above its instrumental value for my preventing as many suffering moments as I can?
I don’t see why {resolving conflicts by weighing everything (ultimately) in suffering} would ever lead me to {“accept suffering to avoid happiness”}, if happiness already can be weighed against suffering in terms of its suffering-preventing effects—just not by itself, which is what many other utilitarianisms rely on, inviting grotesque problems like doctors having parties so great that they outweigh the untreated suffering of their patients.
Are there also practical situations where I’d want to admit that paperclips have terminal value, or else accept suffering to avoid paperclips?
I don’t see what hidden assumptions I’m missing here. I certainly don’t think an infinitely large paperclip is an acceptable comparand to outweigh any kind of suffering. In the case of happiness, it depends completely on whether the combined causal cascades from this happiness are expected to prevent more suffering than the current comparand suffering: no need to attach any independent numerical terminal value to happiness itself, or we’d be back to counting happy sheep believing it to outweigh someone’s agony any moment now.
Also NU in general has a bad reputation in the philosophy community (more than classical utilitarianism I think) so it's better EAs don't endorse it.
I believe the first part of this statement may currently be true for the WEIRD (western, educated, industrialized, rich, democratic) philosophy community. Other parts of the world have long histories and living traditions of suffering-based views, primarily various forms of Buddhism. In what I’ve read about Mahayana Buddhism (or the Bodhisattva path), compassion is often explicitly identified as the only necessary motivation that implies and/or transcends all the outwardly visible customs, rules, and ethics, and that compassion is the voice to listen to when other “absolutes” conflict. (Omnicidal superweapon research is not part of these philosophies of compassion, but invented, in my estimation, as an implication of NU by later armchair rationalists to easily dismiss NU.)
I’ll take the second part of your statement as your current personal opinion of NU in its present form and perceived reputation. I am personally still optimistic that suffering is the most universal candidate to derive all other values from, and I would be careful not to alienate a large segment of systematic altruists such as might be found among secular, rationalist Buddhists. I mostly agree though, that NU in its present form may be tainted by the prevalence of the world-destruction argument (even though it is argued to represent only a straw man NU by proponents of NU).
9 comments
Comments sorted by top scores.
comment by ESRogs · 2019-02-14T19:27:58.309Z · LW(p) · GW(p)
Meta-note: I am surprised by the current karma rating of this question. At present, it is sitting at +9 points with 7 votes, but it would be at +2 with 6 votes w/o my strong upvote.
To those who downvoted, or do not feel inclined to upvote -- does this question not seem like a good use of LW's question system? To me it seems entirely on-topic, and very much the kind of thing I would want to see here. I found myself disagreeing with much of the text, but it seemed to be an honest question, sincerely asked.
Was it something about the wording (either of the headline or the explanatory text) that put you off?
Replies from: Dagon↑ comment by Dagon · 2019-02-15T16:21:28.094Z · LW(p) · GW(p)
I didn't vote - it seems appropriate for LW, but didn't have any links or references, so it seemed to be a reaction to an argument that I wasn't familiar with. The bold text (only one line, so presumably the main point of the post) was dismissing something as a strawman without explaining why.
It did spawn some interesting conversation, so I've now upvoted it.
comment by Raemon · 2019-02-12T00:42:30.427Z · LW(p) · GW(p)
Serious question: are you depressed? So far most negative utilitarians I've known were depressed and some stopped being negative utilitarian after fixing a chemical imbalance that hampered their ability to experience good things.
Replies from: Teo Ajantaival↑ comment by Teo Ajantaival · 2019-02-12T10:41:17.777Z · LW(p) · GW(p)
No, I’m not depressed, and I believe I never have been. I understand and appreciate the question if what you describe is your prior experience of people who identify as negative utilitarians. I may identify as NU for discussion’s sake, but my underlying identification is with the motivation of impartial compassion. I would go as far as to say that I am happy in all areas of my personal life, being driven towards unification by my terminal concern for the expected suffering of others.
I have had brief experiences of medical emergencies that gave me new perspectives into suffering from the inside. (In hindsight, much of it was generated by fear and escalating perception, but so it is in real danger.) While those happened years ago, I’ve continued to reflect on them and feel like they’ve changed me, affecting my daily life and priorities since then. For a while, I felt grateful for getting my life back and considered devoting myself to treating acute pain. I since graduated Master’s in Psychology without clinical internship to focus more on research, feeling that my comparative advantage is in channeling compassion for more scalable, theoretical work.
I believe a possible mistake of depressed NUs is to focus on others’ suffering before taking care of themselves (by listening to the foundational motivation of self-compassion). Self-compassion is our value-grounding for extended self-compassion, which leads to universal compassion in the limit.
Nate Soares has a post about self-compassion as a key part of his wider, 40-post series, Replacing Guilt, both of which I universally recommend (also ePUB-compiled here [LW · GW]).
Replies from: Raemoncomment by ioannes (ioannes_shade) · 2019-02-15T16:58:18.887Z · LW(p) · GW(p)
I like tranquilism as a view preserving some of the attractive aspects of negative utilitarianism whilst sidestepping some of its icky implications.
comment by Dagon · 2019-02-12T17:04:30.876Z · LW(p) · GW(p)
Can you provide a pointer or summary to how ANU doesn't require commensurability of values? How does one evaluate reduction in suffering of many small-experience/short-lived creatures at the expense of increased suffering of a deeper-experiencing/longer-lived creature?
Also, how does ANU (or any system trying to minimize it's only metric) deal with population-size issues? Should Thanos snap his fingers a lot more often (removing suffering by removing entities, painlessly and by surprise)?
I'll admit that neither of these are my true objection. My felt experience and strong intuition is that suffering can coexist with joy and satisfaction, and one should not sacrifice much of the latter in order to reduce the former.
comment by aogara (Aidan O'Gara) · 2019-02-11T23:52:32.848Z · LW(p) · GW(p)
I haven't seriously evaluated the arguments, but my intuition is that suffering and happiness are opposite sides of the same scale, not separate values. Utility is the measure of how good or bad something is, and happiness and suffering correspond to positive and negative values of utility.
terminal value monism (suffering is the only thing that motivates us “by itself”)
So I'd say that I value only utility.
it is a straw man argument that NUs don’t value life or positive states, because NUs value them instrumentally
Again, not having thought too much about it, I find that my intuition better matches a system that cares about positive utility even when it doesn't avert negative utility. E.g., I want paradise forever, not mild pleasantness forever.
Is there a good reason to suspect this is wrong?
comment by educapaverde · 2022-06-16T22:40:11.376Z · LW(p) · GW(p)
Avoiding suffering isn't terminally valuable, but instrumentally valuable.