Posts

LLM+Planners hybridisation for friendly AGI 2024-05-03T08:40:18.128Z
Empathy bandaid for immediate AI catastrophe 2023-04-05T02:12:54.713Z

Comments

Comment by installgentoo on LLM+Planners hybridisation for friendly AGI · 2024-05-03T22:42:04.616Z · LW · GW

Do you want to make a demo with dspy+gpt4 api+fast downward?

Comment by installgentoo on But why would the AI kill us? · 2023-04-18T12:32:32.084Z · LW · GW

You can make AI care about us with this one weird trick:

1. Train a separate agent action reasoning network. For LLM tech this should be training on completing interaction sentences, think "Alice pushed Bob. ___ fell due to ___", with a tokenizer that generalizes agents(Alice and Bob) into generic {agent 1, agent n} and "self agent". Then we replace various Alices and Bobs in various action sentences with generic agent tokens, and train on guessing consequences or prerequisites of various actions from real situations that you can get from any text corpus.

2. Prune anything that has to do with agent reasoning from the parent LLM. Any reasoning has to go through the agent reasoning network.

3. In anything that has to do with Q-learning cost we replace tokens in agent reasoning net with "self token", rerun the network, and take highest cost of two. Repeat this for every agent token. Agent will be forced to treat harm to anyone else as harm to itself.

Emulated empathy. I'm concerned that this might produce a basilisk though, also it will be inherently weak against predatory ai as it will be forced to account for its wellbeing too. Might work though if these agents are deployed autonomously and allowed to exponentially grow. I dunno. I think we'll all die this year, knowing humanity.