Posts

An argument that consequentialism is incomplete 2024-10-07T09:45:12.754Z
Population ethics and the value of variety 2024-06-23T10:42:21.402Z
Book review: The Quincunx 2024-06-05T21:13:55.055Z
A case for fairness-enforcing irrational behavior 2024-05-16T09:41:30.660Z
I'm open for projects (sort of) 2024-04-18T18:05:01.395Z
A short dialogue on comparability of values 2023-12-20T14:08:29.650Z
Bounded surprise exam paradox 2023-06-26T08:37:47.582Z
Stop pushing the bus 2023-03-31T13:03:45.543Z
Aligned AI as a wrapper around an LLM 2023-03-25T15:58:41.361Z
Are extrapolation-based AIs alignable? 2023-03-24T15:55:07.236Z
Nonspecific discomfort 2021-09-04T14:15:22.636Z
Fixing the arbitrariness of game depth 2021-07-17T12:37:11.669Z
Feedback calibration 2021-03-15T14:24:44.244Z
Three more stories about causation 2020-11-03T15:51:58.820Z
cousin_it's Shortform 2019-10-26T17:37:44.390Z
Announcement: AI alignment prize round 4 winners 2019-01-20T14:46:47.912Z
Announcement: AI alignment prize round 3 winners and next round 2018-07-15T07:40:20.507Z
How to formalize predictors 2018-06-28T13:08:11.549Z
UDT can learn anthropic probabilities 2018-06-24T18:04:37.262Z
Using the universal prior for logical uncertainty 2018-06-16T14:11:27.000Z
Understanding is translation 2018-05-28T13:56:11.903Z
Announcement: AI alignment prize round 2 winners and next round 2018-04-16T03:08:20.412Z
Using the universal prior for logical uncertainty (retracted) 2018-02-28T13:07:23.644Z
UDT as a Nash Equilibrium 2018-02-06T14:08:30.211Z
Beware arguments from possibility 2018-02-03T10:21:12.914Z
An experiment 2018-01-31T12:20:25.248Z
Biological humans and the rising tide of AI 2018-01-29T16:04:54.749Z
A simpler way to think about positive test bias 2018-01-22T09:38:03.535Z
How the LW2.0 front page could be better at incentivizing good content 2018-01-21T16:11:17.092Z
Beware of black boxes in AI alignment research 2018-01-18T15:07:08.461Z
Announcement: AI alignment prize winners and next round 2018-01-15T14:33:59.892Z
Announcing the AI Alignment Prize 2017-11-04T11:44:19.000Z
Announcing the AI Alignment Prize 2017-11-03T15:47:00.092Z
Announcing the AI Alignment Prize 2017-11-03T15:45:14.810Z
The Limits of Correctness, by Bryan Cantwell Smith [pdf] 2017-08-25T11:36:38.585Z
Using modal fixed points to formalize logical causality 2017-08-24T14:33:09.000Z
Against lone wolf self-improvement 2017-07-07T15:31:46.908Z
Steelmanning the Chinese Room Argument 2017-07-06T09:37:06.760Z
A cheating approach to the tiling agents problem 2017-06-30T13:56:46.000Z
What useless things did you understand recently? 2017-06-28T19:32:20.513Z
Self-modification as a game theory problem 2017-06-26T20:47:54.080Z
Loebian cooperation in the tiling agents problem 2017-06-26T14:52:54.000Z
Thought experiment: coarse-grained VR utopia 2017-06-14T08:03:20.276Z
Bet or update: fixing the will-to-wager assumption 2017-06-07T15:03:23.923Z
Overpaying for happiness? 2015-01-01T12:22:31.833Z
A proof of Löb's theorem in Haskell 2014-09-19T13:01:41.032Z
Consistent extrapolated beliefs about math? 2014-09-04T11:32:06.282Z
Hal Finney has just died. 2014-08-28T19:39:51.866Z
"Follow your dreams" as a case study in incorrect thinking 2014-08-20T13:18:02.863Z
Three questions about source code uncertainty 2014-07-24T13:18:01.363Z

Comments

Comment by cousin_it on How to end credentialism · 2025-04-20T22:37:29.044Z · LW · GW
Comment by cousin_it on A Dissent on Honesty · 2025-04-15T11:54:06.455Z · LW · GW

I mean, Flynn Rider was also really good-looking. For a lot of people, maybe most, this look is just unattainable. Even if you can get in as good physical shape (which is far from easy), what if you're older, shorter, balder, have a goofy face and so on.

Comment by cousin_it on What is autism? · 2025-04-12T19:58:28.935Z · LW · GW

Yeah. I had a similar idea, that autism spectrum stuff comes from a person's internal "volume knobs" being turned to the wrong positions. Some things are too quiet to notice, while others are so loud that it turns into a kind of wailing feedback, like from a too loud microphone. And maybe some of it is fixable with exposure training, but not everything and not easily.

Comment by cousin_it on My day in 2035 · 2025-04-11T18:46:17.277Z · LW · GW
Comment by cousin_it on Weird Random Newcomb Problem · 2025-04-11T18:38:35.356Z · LW · GW

Wei's motivating example for UDT1.1 is exactly that. It is indeed weird that Eliezer's FDT paper doesn't use the idea of optimizing over input-output maps, despite coming out later. But anyway, "folklore" (which is slowly being forgotten it seems) does know the proper way to handle this.

Comment by cousin_it on Weird Random Newcomb Problem · 2025-04-11T18:21:33.442Z · LW · GW

I'm not sure question 1 is analogous to Newcomb's problem. We're trying to maximize the money made by the program overall, not just the money made in the specific (x,x) case. In other words, even if you're "inside" the problem, the UDT-ish thing to do is first figure out the optimal way to respond to all (x,y) pairs, and only then feed it the (x,x) pair. Wei Dai called this "UDT1.1", first optimizing over all input-output maps and then applying it to a specific input.

Comment by cousin_it on Quarter Inch Cables are Devious · 2025-04-05T15:46:07.433Z · LW · GW

ideally the only 1/4″ cables onstage are short runs to DIs

And all the pedalboard stuff that happens before the DI. But mostly I agree.

Btw, do you already know that a piezo signal is much improved by a preamp with >1 meg ohm input impedance? I figured that out with my electric cello.

Comment by cousin_it on AI #110: Of Course You Know… · 2025-04-04T07:57:01.217Z · LW · GW

I think there's a worldwide trend toward more authoritarian leaders, which contributed to both these events. And it should raise our probability of e.g. Turkey or China doing something silly. But where this trend comes from, I'm not sure. It certainly predates the current AI wave. It could be due to social media making people more polarized or something. But then again there were plenty of worse dictators in history, long before social media or electricity. So maybe what's happening now is regression to the mean, and nice democracy was an anomaly in place and time.

Comment by cousin_it on AI #110: Of Course You Know… · 2025-04-03T23:30:06.734Z · LW · GW

Yeah. I remember where I was and how I felt when covid hit in 2020, and when Russia attacked Ukraine in 2022. This tariff announcement was another event in the same row.

And it all seems so stupidly self-inflicted. Russia's economy was booming until Feb 2022, and US economy was doing fine until Feb 2025. Putin-2022 and Trump-2025 would've done better for their countries by simply doing nothing. Maybe this shows the true value of democratic checks and balances: most of the time they add overhead, but sometimes they'll prevent some exceptionally big and stupid decision, and that pays for all the overhead and then some.

Comment by cousin_it on Solving willpower seems easier than solving aging · 2025-03-24T11:23:06.938Z · LW · GW

Your examples sound familiar to me too, but after rereading your comment and mine, maybe it all can be generalized in a different way. Namely, that internal motivation leads to a low level of effort: reading some textbooks now and then, solving some exercises, producing some small things. It still feels a bit like staying in place. Whereas it takes external motivation to actually move forward with math, or art, or whatever - to spend lots of effort and try to raise my level every day. That's how it feels for me. Maybe some people can do it without external motivation, or maybe they lucked into getting external motivation in the right way, I don't know.

Comment by cousin_it on Solving willpower seems easier than solving aging · 2025-03-23T22:30:03.060Z · LW · GW

I agree feedback is a big part of it. For example, the times in my life when I've been most motivated to play musical instruments were when I had regular opportunities to play in front of people. Whenever that disappeared, the interest went away too.

But also I think some of it is sticky, or due to personality factors. We could even say it's not about willpower at all, but about value differences. Some people are just more okay with homeostasis, staying at a certain level (which can be lower or higher for different people) and using only as much effort as needed for that. While others keep climbing and applying effort without ever reaching a level that lets them relax. Many billionaires seem to be of that second type. I'm more of the first type, with many of my active periods being prompted by external changes, threats to homeostasis. It's clear that type 2 achieves more than type 1, but it's not clear which type is happier and whether one should want to switch types.

Comment by cousin_it on Towards a scale-free theory of intelligent agency · 2025-03-22T08:04:11.165Z · LW · GW

Good post. But I thought about this a fair bit and I think I disagree with the main point.

Let's say we talk about two AIs merging. Then the tuple of their expected utilities from the merge had better be on the Pareto frontier, no? Otherwise they'd just do a better merge that gets them onto the frontier. Which specific point on the frontier is a matter of bargaining, but the fact that they want to hit the frontier isn't, it's a win-win. And the merges that get them to the frontier are exactly those that output a EUM agent, maximizing some linear combination of their utilities. If the point they want to hit is in a flat region of the frontier, the merge will involve coinflips to choose which EUM agent to become; and if it's curvy at that point, the merge will be deterministic. For realistic agents who have more complex preferences than just linearly caring about one cake, I expect the frontier will be curvy, so deterministic merge into a EUM agent will be the best choice.

Comment by cousin_it on Elite Coordination via the Consensus of Power · 2025-03-19T09:46:47.585Z · LW · GW

"Apparatchik" in the USSR was some middle-aged Ivan Ivanovich who'd yell at you in his stuffy office for stepping out of line. His power came from the party apparatus. While the power of Western activists is the opposite: it comes from civil society, people freely associating with each other.

This rhetorical move, calling a Western thing by an obscure and poorly fitting Soviet name, is a favorite of Yarvin: "Let's talk about Google, my friends, but let's call it Gosplan for a moment. Humor me." In general I'd advise people to stay away from his nonsense, it's done enough harm already.

Comment by cousin_it on Counting Objections to Housing · 2025-03-17T01:37:19.464Z · LW · GW

The objection I'm most interested in right now is the one about induced demand (that's not the right term but let's roll with it). Like, let's say we build many cheap apartments in Manhattan. Then the first bidders for them will be rich people - from all over the world! - who would love to get a Manhattan apartment for a bargain price. The priced-out locals will stay just as priced out, shuffled to the back of the line, because there's quite many rich people in the world who are willing to outbid them. Maybe if we build very many apartments, and not just in Manhattan but everywhere, the effect will eventually run out; but it'll take very many indeed.

The obvious fix is to put a thumb on the scale somehow, for example sell these cheap apartments only as primary residences. But then we lose the theoretical beauty of "just build more", and we really should figure out what mix of "just build more" and "put a thumb on the scale" is the most cost-efficient for achieving what we want. Maybe some thumb on the scale will even give us what we want without building more, since there's a lot of empty housing and non-primary housing.

Comment by cousin_it on cousin_it's Shortform · 2025-03-07T22:49:22.359Z · LW · GW
Comment by cousin_it on Give Neo a Chance · 2025-03-06T02:08:00.608Z · LW · GW

Maybe you're pushing your proposal a bit much, but anyway as creative writing it's interesting to think about such scenarios. I had a sketch for a weird utopia story where just before the singularity, time stretches out for humans because they're being run at increasing clock speed, and the Earth's surface also becomes much larger and growing. So humanity becomes this huge, fast-running civilization living inside an AI (I called it "Quetzalcoatl", not sure why) and advising it how it should act in the external world.

Comment by cousin_it on The old memories tree · 2025-03-06T00:51:25.064Z · LW · GW

My wife used to have a talking doll that said one phrase in a really annoying voice. Well, at some point the doll short-circuited or something, and started turning on at random times. In the middle of the night for example it would yell out its phrase and wake everyone up. So eventually my wife took the doll to the garbage dump. And on the way back she couldn't stop thinking about the doll sitting there in the garbage, occasionally yelling out its phrase: "Let's go home! I'm already hungry!" This isn't creative writing btw, this actually happened.

Comment by cousin_it on On Writing #1 · 2025-03-04T14:55:53.854Z · LW · GW

The thread about Tolkien reminded me of Andrew Hussie's writing process. Start by writing cool scenes, including any elements you like. A talking tree? Okay. Then worry about connecting it with the story. The talking tree comes from an ancient forest and so on. And if you're good, the finished story will feel like it always needed a talking tree.

I'd be really interested in a similar breakdown of JK Rowling's writing process, because she's another author with a limitless "toybox".

Comment by cousin_it on Richard_Kennaway's Shortform · 2025-03-03T19:20:17.661Z · LW · GW

I think something like the Culture, with aligned superintelligent "ships" keeping humans as basically pets, wouldn't be too bad. The ships would try to have thriving human societies, but that doesn't mean granting all wishes - you don't grant all wishes of your cat after all. Also it would be nice if there was an option to increase intelligence, conditioned on increasing alignment at the same time, so you'd be able to move up the spectrum from human to ship.

Comment by cousin_it on Historiographical Compressions: Renaissance as An Example · 2025-03-03T13:11:51.220Z · LW · GW
Comment by cousin_it on Self-fulfilling misalignment data might be poisoning our AI models · 2025-03-03T00:38:36.448Z · LW · GW
Comment by cousin_it on Osaka · 2025-03-02T01:58:43.282Z · LW · GW

Maybe tangential, but this reminded me of a fun fact about Hong Kong's metro: it's funded by land value. They put a station and get some land development rights near it. Well, building the station obviously makes land around it more valuable. So they end up putting stations where they'd be most useful, and fares can be cheap because the metro company makes plenty of money from land. So the end result is cheap, well-planned public transport which is profitable and doesn't take government money.

Comment by cousin_it on Historiographical Compressions: Renaissance as An Example · 2025-03-02T00:58:29.314Z · LW · GW

Not to pick on you specifically, but just as a general comment, I'm getting a bit worried about the rationalist book review pipeline. It seems it usually goes like this: someone writes a book with an interesting idea -> a rationalist (like Scott) writes a review of it, maybe not knowing much about the topic but being intrigued by the idea -> lots of other rationalists get the idea cached in their minds. So maybe it'd be better if book reviews were written by people who know a lot about the topic, and can evaluate the book in context.

Like, a while ago someone on LW asked people to recommend textbooks on various topics, but you couldn't recommend a textbook if it was the only one you'd read on the topic, you had to read at least two and then recommend one. That seems on the right track to me, and requiring more knowledge of the topic would be better still.

Comment by cousin_it on For the Sake of Pleasure Alone · 2025-02-28T13:34:08.307Z · LW · GW

I think I can destroy this philosophy in two kicks.

Kick 1: pleasure is not one-dimensional. There are different parts of your brain that experience different pleasures, with no built-in way to compare between them.

When you retreat from kick 1 by saying "my decision-making provides a way to compare, the better pleasure is the one I'll choose when asked", here comes kick 2: your decision-making won't work for that. There are compulsive behaviors that people want to do but don't get much pleasure from them. And in every decision there's a possible component of that, however small.

You could say "I'll compare decisions based on how much pleasure they bring, excluding compulsiveness", but you can't do that due to kick 1 again. So the philosophy just collapses.

Comment by cousin_it on How to Make Superbabies · 2025-02-27T23:20:05.422Z · LW · GW

Good point. But I think the real game changer will be self-modification tech, not longevity tech. In that case we won't have a "slow adaptation" problem, but we'll have a "fast adaptation in weird directions" problem which is probably worse.

Comment by cousin_it on Osaka · 2025-02-26T15:01:54.889Z · LW · GW

In Copenhagen every street has wide sidewalks and bike lanes in both directions, and there's lots of public transport too. It's good.

Comment by cousin_it on Export Surplusses · 2025-02-24T12:03:50.034Z · LW · GW

I don't understand Eliezer's explanation. Imagine Alice is hard-working and Bob is lazy. Then Alice can make goods and sell them to Bob. Half the money she'll spend on having fun, the other half she'll save. In this situation she's rich and has a trade surplus, but the other parts of the explanation - different productivity between different parts of Alice (?) and inability to judge her own work fairly (?) - don't seem to be present.

Comment by cousin_it on The case for the death penalty · 2025-02-22T21:57:30.182Z · LW · GW
Comment by cousin_it on The case for the death penalty · 2025-02-22T01:23:09.193Z · LW · GW
Comment by cousin_it on The case for the death penalty · 2025-02-22T01:07:14.269Z · LW · GW

No. Committing a crime inflicts damage. But interacting with a person who committed a crime in the past doesn't inflict any damage on you.

Comment by cousin_it on The case for the death penalty · 2025-02-21T23:36:32.340Z · LW · GW

Because the smaller measure should (on my hypothesis) be enough to prevent crime, and inflicting more damage than necessary for that is evil.

Comment by cousin_it on The case for the death penalty · 2025-02-21T22:32:20.913Z · LW · GW

Because otherwise everyone will gleefully discriminate against them in every way they possibly can.

Comment by cousin_it on The case for the death penalty · 2025-02-21T19:06:49.758Z · LW · GW

I think the US has too much punishment as it is, with very high incarceration rate and prison conditions sometimes approaching torture (prison rape, supermax isolation).

I'd rather give serial criminals some kind of surveillance collars that would detect reoffending and notify the police. I think a lot of such people can be "cured" by high certainty of being caught, not by severity of punishment. There'd need to be laws to prevent discrimination against people with collars, though.

Comment by cousin_it on Ascetic hedonism · 2025-02-17T20:46:18.788Z · LW · GW

Yeah, I stumbled on this idea a long time ago as well. I never drink sugary drinks, my laptop is permanently in grayscale mode and so on. And it doesn't feel like missing out on fun; on the contrary, it allows me to not miss out. When I "mute" some big, addictive, one-dimensional thing, I start noticing all the smaller things that were being drowned out by it. Like, as you say, noticing the deliciousness of baked potatoes when you're not eating sugar every day, or noticing all the colors in my home and neighborhood when my screen is on grayscale.

Comment by cousin_it on Celtic Knots on a hex lattice · 2025-02-15T00:59:08.713Z · LW · GW
Comment by cousin_it on Altman blog on post-AGI world · 2025-02-10T09:57:30.008Z · LW · GW

I suppose the superassistants could form coalitions and end up as a kind of "society" without too much aggression. But this all seems moot, because superassistants will anyway get outcompeted by AIs that focus on growth. That's the real danger.

Comment by cousin_it on Altman blog on post-AGI world · 2025-02-10T01:15:06.910Z · LW · GW

I don't quite understand the plan. What if I get access to cheap friendly AI, but there's also another much more powerful AI that wants my resources and doesn't care much about me? What would stop the much more powerful AI from outplaying me for these resources, maybe by entirely legal means? Or is the idea that somehow the AIs in public access are always the strongest possible? That isn't true even now.

Comment by cousin_it on The Risk of Gradual Disempowerment from AI · 2025-02-06T00:45:57.299Z · LW · GW

I also agree with all of this.

For what an okayish possible future could look like, I have two stories in mind:

  1. Humans end up as housecats. Living among much more powerful creatures doing incomprehensible things, but still mostly cared for.

  2. Some humans get uplifted to various levels, others stay baseline. The higher you go, the more aligned you must be to those below. So still a hierarchy, with super-smart creatures at the top and housecats at the bottom, but with more levels in between.

A post-AI world where baseline humans are anything more than housecats seems hard to imagine, I'm afraid. And even getting to be housecats at all (rather than dodos) looks to be really difficult.

Comment by cousin_it on Tear Down the Burren · 2025-02-04T10:59:31.295Z · LW · GW

Thanks for writing this, it's a great explanation-by-example of the entire housing crisis.

Comment by cousin_it on Predation as Payment for Criticism · 2025-02-02T23:57:27.926Z · LW · GW

Well, Christianity sometimes spread by conquest, but other times it spread peacefully just as effectively. Same for democracy. So I don't think the spread of moral values requires conquest.

Comment by cousin_it on The Simplest Good · 2025-02-02T23:41:02.226Z · LW · GW

Wait, but we know that people sometimes have happy moments. Is the idea that such moments are always outweighed by suffering elsewhere? It seems more likely that increasing the proportion of happy moments is doable, an engineering problem. So basically I'd be very happy to see a world such as in the first half of your story, and don't think it would lead to the second half.

Comment by cousin_it on Gradual Disempowerment, Shell Games and Flinches · 2025-02-02T15:28:15.377Z · LW · GW
Comment by cousin_it on Poetic Methods I: Meter as Communication Protocol · 2025-02-01T21:47:55.969Z · LW · GW
Comment by cousin_it on Predation as Payment for Criticism · 2025-01-31T18:13:23.067Z · LW · GW

Your theory would predict that we'd be much better at modeling tigers (which hunted us) than at modeling antelopes (which we hunted), but in reality we're about equally bad at modeling either, and much better at modeling other humans.

Comment by cousin_it on The future of humanity is in management · 2025-01-30T23:08:28.423Z · LW · GW

I don't think this post addresses the main problem. Consider the exchange ratio between labor and land. You need land to live, and your food needs land to be grown. Will you be able to afford more land use for the same work hours, or less? (As programmer, manager, CEO, super high productivity job, whatever.) Well, if the same land can be used to run AIs that can do your job N times over, then from your labor you won't be able to afford it, and that closes the case.

So basically, the only way the masses can survive long term is by some kind of handouts. It won't just happen by itself due to tech progress and economic laws.

Comment by cousin_it on Predation as Payment for Criticism · 2025-01-30T02:12:19.326Z · LW · GW

I don't buy it. Lots of species have predators and have had them for a long time, but very few species have intelligence. It seems more likely that most of our intelligence is due to sexual selection, a Fisherian runaway that accidentally focused on intelligence instead of brightly colored tails or something.

Comment by cousin_it on The Manhattan Trap: Why a Race to Artificial Superintelligence is Self-Defeating · 2025-01-22T08:49:06.505Z · LW · GW

An ASI project would be highly distinguishable from civilian AI applications and not integrated with a state’s economy

Why? I think there's a smooth ramp from economically useful AI to superintelligence: AIs gradually become better at many tasks, and these tasks help more and more with improving AI in turn.

Comment by cousin_it on Tax Price Gouging? · 2025-01-18T16:58:03.057Z · LW · GW
Comment by cousin_it on We probably won't just play status games with each other after AGI · 2025-01-16T18:47:42.723Z · LW · GW
Comment by cousin_it on RobertM's Shortform · 2025-01-16T14:08:54.685Z · LW · GW

For cognitive enhancement, maybe we could have a system like "the smarter you are, the more aligned you must be to those less smart than you"? So enhancement would be available, but would make you less free in some ways.