[Link] “Proxy measures, sunk costs, and Chesterton's fence”, or: the sunk cost heuristic

post by kpreid · 2012-08-08T14:39:02.580Z · LW · GW · Legacy · 5 comments

Contents

5 comments

Thought this post might be of interest to LW: Proxy measures, sunk costs, and Chesterton's fence. To summarize: Previous costs are a proxy measure for previous estimates of value, which may have information current estimates of value do not; therefore acting according to the sunk cost fallacy is not necessarily wrong.

This is not an entirely new idea here, but I liked the writeup. Previous discussion: Sunk Costs Fallacy Fallacy; Is Sunk Cost Fallacy a Fallacy?.

Excerpt:

If your evidence may be substantially incomplete you shouldn't just ignore sunk costs — they contain valuable information about decisions you or others made in the past, perhaps after much greater thought or access to evidence than that of which you are currently capable. Even more generally, you should be loss averse — you should tend to prefer avoiding losses over acquiring seemingly equivalent gains, and you should be divestiture averse (i.e. exhibit endowment effects) — you should tend to prefer what you already have to what you might trade it for — in both cases to the extent your ability to measure the value of the two items is incomplete. Since usually in the real world, and to an even greater degree in our ancestors' evolutionary environments, our ability to measure value is and was woefully incomplete, it should come as no surprise that people often value sunk costs, are loss averse, and exhibit endowment effects — and indeed under such circumstances of incomplete value measurement it hardly constitutes "fallacy" or "bias" to do so.

5 comments

Comments sorted by top scores.

comment by OrphanWilde · 2012-08-08T15:12:21.297Z · LW(p) · GW(p)

While I agree with the post, I'm not sure that it's actually a strong defense of sunk costs as being more heuristic than fallacy; it depends upon your past self having more information than your current self. I believe the sunk cost fallacy is generally used to refer to the phenomenon where, having additional information that makes the original investment look like a bad idea, you proceed to additional investment instead of cutting loose.

That is, the sunk cost fallacy generally refers to a situation in which it is more or less explicitly stated that you have more information, rather than less. Starting from the assumption that you have less information in the future than in the past permits this reasoning, but the question becomes whether or not the assumption actually holds. My intuitive reaction is that it doesn't.

Replies from: billswift
comment by billswift · 2012-08-08T17:31:05.558Z · LW(p) · GW(p)

it depends upon your past self having more information than your current self.

Or maybe you just spent more time thinking it through before. "Never doubt under pressure what you have calculated at leisure." I think that previous states should have some influence on your current choices. As the link says:

If your evidence may be substantially incomplete you shouldn't just ignore sunk costs -- they contain valuable information about decisions you or others made in the past, perhaps after much greater thought or access to evidence than that of which you are currently capable.

Replies from: drethelin, OrphanWilde
comment by drethelin · 2012-08-08T19:46:12.729Z · LW(p) · GW(p)

Also remember the corollary that any decision made under pressure could probably stand to be reviewed at leisure.

Replies from: billswift
comment by billswift · 2012-08-08T23:05:55.121Z · LW(p) · GW(p)

Everybody does that anyway, it is usually called second-guessing yourself. The best rule is to not decide under pressure unless you really have to, take the time to think things through.

comment by OrphanWilde · 2012-08-08T19:58:50.128Z · LW(p) · GW(p)

That presumes you've forgotten why you did something to begin with, your reasoning having created that information. Again, given the precise conditions, I think it's a perfectly fine argument. I just don't find those conditions more probable than the converse, which is to say, having more information.