Instrumental ignoring AI, Dumb but not useless.
post by Donald Hobson (donald-hobson) · 2022-10-30T16:55:47.555Z · LW · GW · 6 commentsContents
6 comments
This post is yet another unworkable agent design that uses infinite compute. It outlines an AI that ignores all instrumental values, pursuing only it's terminal values.
What is does it look like to pursue an instrumental value. It means you can predict the AI is likely to do action X, even if you have no idea what it's utility function is.
This AI doesn't do that. It obeys the simple principle that, if you are clueless about it's utility function, then you are clueless about it's actions.
Let mean the probability distribution over the set .
Lets have some set of Observations , Actions and (bounded? ) Utilities .
Then Pick some probability distribution over utility functions , such as uniform or complexity weighted. And some probability distribution over the action space.
Then let the AI be a function
We can formalize our condition as .
(Note refers to the probability that assigns to .)
Why might we expect such agents to be safe. Because if we picked a random utility function, we get a safely random output. So all humans need to do is be better than random in our choice of utility function.
Now let be the expected utility. Choose to maximize subject to the condition. As the condition doesn't relate cases with different values of , this optimization only needs to be computed for the particular observed.
Is this AI design safe when repeatedly called with the same utility function?
Imagine the action space just consists of outputting 0 or 1. And there are only 2 utility functions, paperclips and staples. We could imagine that when run millions of times with a utility of paperclips, this AI outputs code for a paperclip maximizer, and when run millions of times with a goal of staples, it outputs a bitwise negation of the paperclip maximizer. (Which could possibly be a staple maximizer, if the AI has done some op code magic)
This would be because the humans choice of the same utility function each time is worse than random. I think this is an actual problem. At least if the AI's have total knowledge of past and future, then this is one Schelling point to their coordination game. (The consider all possibilities doesn't quite work well with multiple different AI's)
I have an intuition that something TDT'ish might fix this.
The other problem that comes to mind is that it isn't easy to whitelist a limited set of instrumental actions you are ok with. You can fake it though, suppose you have a robot, instead of piping the output of this AI directly into the motors, you have it call simple hardcoded functions to walk and pick up objects. The result being, if the AI has a random utility function, instead of thrashing randomly on the floor, it walks to random places and picks up random objects. Because someone hardcoded those low level commands into the robot.
6 comments
Comments sorted by top scores.
comment by Gurkenglas · 2022-10-30T19:08:59.446Z · LW(p) · GW(p)
Your math can be terser: f:U->ΔA. g:ΔU. h:ΔA. The condition is: f and g combine into h. To fix your first problem, make A policies, not actions.
My biggest problem here is that f depends on how we represent g since U identifies u with 2u. Silence this warning by normalizing U, and f depends on how we normalize.
Replies from: Gurkenglas↑ comment by Gurkenglas · 2022-10-30T19:17:13.448Z · LW(p) · GW(p)
Ah, we've seen my problem before: Solve bargaining [? · GW], then make g bargain to choose f.
comment by TekhneMakre · 2022-10-30T17:22:05.495Z · LW(p) · GW(p)
This seems in danger of being a "sponge alignment" proposal, i.e. the proposed system doesn't do anything useful. https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities#:~:text=sponge [LW · GW]
Replies from: donald-hobson, Gurkenglas↑ comment by Donald Hobson (donald-hobson) · 2022-10-31T01:53:14.801Z · LW(p) · GW(p)
This current version is dumb, but still exerts some optimization pressure. (Just the bits of optimization out are at most the bits of selection put into its utility function.)
↑ comment by Gurkenglas · 2022-10-30T19:12:44.529Z · LW(p) · GW(p)
It could be a conceptual ingredient to something useful. For example, it can select between two plans.
Replies from: TekhneMakre↑ comment by TekhneMakre · 2022-10-31T02:26:19.426Z · LW(p) · GW(p)
I agree.