Roundabout Strategy

post by NicholasKross · 2021-01-28T00:44:00.743Z · LW · GW · 1 comments

This is a link post for https://www.thinkingmuchbetter.com/main/roundabout-strategy/

Contents

  Roundabout
  The Really Good Example of Henry Ford
  A Big List of Examples
  When Roundabout Goes Wrong-about
  Wait, isn’t this just X?
  Conclusion
None
1 comment

Epistemic status: shilling for Big Concept

In the middle of the forest was a library. In the middle of the library was a book. The book’s ideas were usually not very good. At least, they were pretty debatable.

In the middle of the book was a gem.

Roundabout

The book’s author, Mark Spitznagel, synthesized a bunch of ideas from the history of strategy. He coined a newer name for an ancient concept: roundabout. No, not the polite word for "convoluted" or "that traffic thing". Here, this adjective means "pursuing indirect means, to better achieve your goals." A roundabout strategy is one using those means, rather than directly reaching for the ends.

If you’re inclined towards the ancient [LW · GW] ways of Eastern [LW · GW] philosophy [LW · GW], Spitznagel helpfully references more words for the concept:

"In the Dao, yin (unseen, hidden, passive) balances and is balanced by yang (seen, light, active). The antithesis to shi, the intermediate and circuitous, is li, the immediate and direct. As the "all or nothing" of a pitched battle, li seeks decisive victory in each and every engagement (the "false shortcuts" denounced by the Laozi)…"

"The forward-looking shi strategist takes the roundabout path toward subtle and even intangible intermediate steps, while the li strategist concentrates on the current step, the visible power, the direct, obvious route toward a tangible desired end, relying on force to decide the outcome of every battle. Simply stated, li goes for the immediate hit, while shi seeks first the positional advantage of the setup."

If this sounds hokey or weirdly trivial, an example will make the point clearer. Spitznagel’s canonical roundabout story is that of Henry Ford.

The Really Good Example of Henry Ford

In the early days of automobiles, a workshop of skilled workers would build each car one at a time. This is the direct, or li, way to get cars. To become the largest auto manufacturer, Ford picked a different route.

He built the River Rouge factory, optimized to build consistent-quality cars at a high speed. An assembly line took the task of car-building, split it into microsteps, and optimized each one. Cars were not merely assembled, but their component parts were built in-house, and the components of those components. This extended to the raw materials when Ford bought rubber farms and iron mines. The goal: produce cars faster (in the long run) and make them cheaper (on average). The factory started by producing military equipment and parts, before eventually helping build the Model A and Model T cars.

Now, this was a very roundabout way to build cars, compared to just buying some parts and assembling them by hand. The key is the pursuit of indirect means. The small workshops made more cars than River Rouge, because they had a head start. River Rouge was building no cars (and in fact costing tons of money), for several years. Right until it opened, that is. After that, the direct path’s "head start" was obliterated.

Example image

A Big List of Examples

When Roundabout Goes Wrong-about

The indirect route is not always best. You can’t prepare forever; you must eventually attack. And you still have to keep track of the short term, lest something closer in time block you from moving further.

Wait, isn’t this just X?

The point of using the word "roundabout" is to give a quick name to a specific concept. The concept is linked to other key ideas in the rationalist space, but it is not identical to them.

Is "roundabout" a fancy rewording of instrumental goals [LW · GW], as opposed to terminal goals? Well, no. If your terminal goal is personal happiness, one instrumental goal might be making money. Recall the example above: you could work a job or start a business to make money. Starting a businesses is a more roundabout strategy than working a job, but they both work towards an instrumental goal (money). An instrumental goal could be more or less roundabout than another, but making something more roundabout won’t always make something more or less instrumental.

Is "roundabout" another word for long-term thinking, as opposed to focusing on the short term? Well, no. Roundabout strategy is a strategy, not a goal. You can certainly have long-term goals and work towards them, but that isn’t roundabout own its own. A relevant example is the field of longtermist ethics. You could directly work on long term threats to humanity, or write about the philosophical ideas needed to guide the far future. These are both obviously longtermist in their final goals, and one might be more effective than the other. But the second method is more indirect, and thus more roundabout.

Is "roundabout" simply the delayed gratification thing all over again? Well, no. At some point, you have to turn the factory on, make the profit, solve the riddle, and cash out. And, as with the entrepreneurship example, there are risks with longer time horizons and risky plans. Perhaps the one-marshmallow children, in the classic experiment, didn’t think waiting was worth the risk.

Conclusion

The roundabout concept is a useful abstraction for discourse. It intuitively combines planning, risk, and time, and creates a focal point of discussion. Just having the concept on the table suggests more ideas to be thought of. Think back to the Effective Altruism examples. Maybe some EA projects so capital-intensive that, like Ford’s factory, you need to take on risk and sink lots of capital and time. Otherwise, the goal wouldn’t even be possible. With the "roundabout" concept, these discussions can be more explicit, and people’s stances clarified. "I think your approach is too roundabout, because you are underestimating the risks of your long-term plan."

Naming a concept creates a chunk of the concept. "Roundabout" is quicker to remember than "let’s build this thing to help us do the other thing better". It becomes a tool in your thinking, a lens for judging ideas, and a building block of larger plans. My hope is for the roundabout concept to take a seat at the table of decision-making. And, unlike some concepts, roundabout strategy offers ways to create new strategies, instead of "merely" ruling out the bad ones. "How do we trade stocks faster? Well, what’s a roundabout way to get the orders in… maybe a light-speed cable?"


  1. The entire field of quant finance is known for being cartoonishly roundabout. If you want to make money trading stocks, a real roundabout strategy is to have a fiber-optic cable running to the exchange. It costs 9 figures in the short term, but gives a nice speed buff later. The game of finance certainly has the best example of power creep [LW · GW] in history. ↩︎

  2. And by "people", I really mean "all potential future morally relevant entities". So animals and ems [LW · GW] would be weighed. ↩︎

  3. I saw a diagram showing this link between optimization and abstraction. It was in an article explaining this; small gains at the "top" of the abstraction pyramid, cascade into big gains at the bottom. I don’t remember where I saw this, but I will edit this article to link it, if someone can find it. ↩︎

  4. I’m afraid this will happen with me, in learning the skills for A.I. safety. What if I don’t have a good way to contribute? What if I do, but I don’t contribute in time [LW · GW]? The good news is that I haven’t gotten distracted by a large roundabout plan. "That’s my secret, captain: I’m always distracted..." [LW · GW] ↩︎

1 comments

Comments sorted by top scores.

comment by NicholasKross · 2023-07-12T23:37:29.326Z · LW(p) · GW(p)

To clarify the "Bitter Lesson" example: the non-roundabout "direct" AI strategy is to use the "sweet shortcut" (h/t Gwern), by using existing human expert knowledge and trying to encode that into a computer. The roundabout strategy is to build a massive computing infrastructure first, which scaling requires. Even if no single group actually executed a strategy of "invent better computers and then do ML on them", society as-a-whole did via the compute overhang.