Posts

elbow921's Shortform 2023-07-08T01:34:59.903Z
Veganism and Acausal Trade 2023-03-03T17:44:46.483Z

Comments

Comment by elbow921 (elbow921@aol.com) on Hammertime Day 6: Mantras · 2024-03-15T02:51:09.486Z · LW · GW

I like the mantra, "If we choose to give more effort today, then we are sure to go beyond our past mistakes." This mantra is on my desktop screen.

Comment by elbow921 (elbow921@aol.com) on elbow921's Shortform · 2024-01-06T23:10:18.920Z · LW · GW

Hedonic Treadmill and the Economy

The hedonic treadmill is when permanent changes to living conditions lead to only temporary increases in happiness. This keeps us always wanting improvements to our lives. We often spend money on the newest Iphones and focus our attention on improving our external circumstances. We ignore the quote:

"What lies before us and what lies behind us are tiny matters compared to what lies within us" 

Some people eat chips to quell their boredom. The hedonic treadmill ensures that, despite improvements in income, people are not satisfied. I was surprised by how much the hedonic treadmill dovetails with profit maximization. If they maximized profit, I suspect companies would pay Big Pharma billions not to release drugs that improve the hedonic set point. The antidepressant drugs Big Pharma releases act as mood flatteners, according to https://www.hedweb.com/. 

Comment by elbow921 (elbow921@aol.com) on Combining individual preference utility functions · 2023-08-10T11:50:29.519Z · LW · GW

I have an idea for a possible utility function combination method. It basically normalizes based on how much utility is at stake in a random dictatorship. The combined utility function has these nice properties:

Pareto-optimality wrt all input utilities on all lotteries

Adding Pareto-dominated options (threats) does not change players' utilities

Invariance to utility scaling

Invariance to cloning every utility function

Threat resistance

 

The combination method goes like this:

X=list of utility functions to combine

dist(U)=worlds where random utility function is argmaxed with ties broken to argmax U

Step 1: For each U in X, U=U-expect(argmax U)

Step 2: For each U in X,

U=-U/U(null universe) if (expected value of U on dist(U))==0

U=-U/(expected value of U on dist(U)) otherwise

Step 3: U(final)=sum(U in X)

 

I designed this to be a CEV of voting utility maximizers, but the combination has multiple discontinuities. It does not need transferrable utility like the ROSE values, but it does not have nearly have as many nice properties as the ROSE values.

Comment by elbow921 (elbow921@aol.com) on Time and Effort Discounting · 2023-08-05T18:38:33.367Z · LW · GW

For the examples in this article, for each option only take the monetary value that goes last. log(amount after year)~0.79*log(amount now)+0.79 is the indifference curve. If U(now)=log(amount now), U(year)=(log(amount after year)-0.79)/0.79.

Comment by elbow921 (elbow921@aol.com) on elbow921's Shortform · 2023-07-08T01:35:00.005Z · LW · GW

Why FAI will not be an expected utility maximizer

Say we have a powerful superintelligent utility maximizer. They will turn the world into the precise configuration that maximizes their expected utility. No human has any say in what will happen.[1]

We do not want our lives optimized for us. We want autonomy, which expected utility maximizers would not give. Nobody has found an outer aligned utility function because powerful expected utility maximizers leave us no room to optimize. Autonomy is one value necessary for futures that we value. Walden One is a dystopia where everyone is secretly manipulated but live happy social lives.

Another reason we hate powerful optimization is status quo bias. Our world is extremely complex, and almost utility functions have maximums far from the current world. This is another reason expected utility maximizers create futures we hate.

We should instead focus on tools that give us an epistesmic advantage and help us choose the world we want to live in. This could involve oracle AI, CEV, training to reduce cognitive biases, etc. This is why I think we should focus on helping people become more rational or approximating effective altruists instead of focusing on inner aligning agent AI.

 

  1. ^

Unless the utility function includes a brain emulation in a position to sculpt the world by choosing the AI's utility function. I do not expect this to happen in practice.

Comment by elbow921 (elbow921@aol.com) on Bing chat is the AI fire alarm · 2023-06-12T21:03:14.759Z · LW · GW

There is a hypothetical example of simulating a ridiculous number of humans typing text and seeing what fraction of those people that type out the current text type out each next token. In the limit, this approaches the best possible text predictor. This would simulate a lot of consciousness.

Comment by elbow921 (elbow921@aol.com) on The Strangest Thing An AI Could Tell You · 2023-05-11T01:04:30.772Z · LW · GW

What if most people would develop superhuman intelligences in their brains without school but, because they have to write essays in school, these superhuman intelligences become aligned with writing essays fast? And no doomsday scenario has happened because they mostly cancel out each others' attempted manipulations and they couldn't program nanobots with their complicated utility functions. ChatGPT writes faster than us and has 20B parameters where humans have 100T parameters, but our neural activations are more noisy than floating-point arithmetic.

Comment by elbow921@aol.com on [deleted post] 2023-04-17T20:06:13.792Z

This is what I am wondering: Does this algorithm, when run, instantiate a subjective experience with the same moral relevance as the subjective experience that happens when mu opioids are released in biological brains?

Comment by elbow921@aol.com on [deleted post] 2023-04-17T15:03:57.502Z

‘By 'obvious to the algorithm' I mean that, to the algorithm, A is referenced with no intermediate computation. This is how pleasure and pain feel to me.  I do not believe all reinforcement learning algorithms feel pleasure/pain. A simple example that does not suffer is the Simpleton iterated prisoner’s dilemma strategy. I believe pain and pleasure are effective ways to implement reinforcement learning. In animals, reinforcement learning is called operant conditioning. See Reinforcement learning on a chicken  for a chicken that has experienced it. I do not know any algorithms to determine whether there is anything to be like a given program. I suspected this program experienced pleasure/pain because of its paralells to the neuroscience of pleasure and pain.

Comment by elbow921@aol.com on [deleted post] 2023-04-17T11:45:41.415Z

As this algorithm executes, the last and 2last variables become the program's last 2 outputs. L1's even indexes become the average input(reward?) given the number of ones the program outputted the last 2 times. I called L1's odd indexes 'confidence' because, as they get higher, the corresponding average reward changes less based on evidence. When L1 becomes entangled with the input generation process, the algorithm chooses which outputs make the inputs higher on average. That is why I called the input 'reward'. L2 reads off the average reward given the last 2 outputs. The algorithm chooses outputs that make the number of ones outputted closer to the number that has yielded the highest inputs in the past. This makes L2 analogous to 'wanting'.