Posts

halwer's Shortform 2025-02-17T18:41:59.534Z

Comments

Comment by halwer on halwer's Shortform · 2025-02-24T17:51:11.716Z · LW · GW

aligning ai beyond human goals: let's say that human goals have a certain effectiveness at increasing the value of the world (or utility). wouldn't we then want ai to improve its own goals to achieve new ones that have increased effectiveness and improving the value of the world? it might not be goals that humans would like, and we might want them to be removed through further alignment, but if they indeed are better we have a duty to allow them.

Comment by halwer on halwer's Shortform · 2025-02-17T15:36:23.847Z · LW · GW

This short piece explores the relationship between human fallibility and the potential role of artificial intelligence in governance. While there is controversy around the nature of AI autonomy, the text presents an argument based on patterns of human leadership and human limitations. I'm particularly interested in hearing perspectives on whether AI could truly transcend human biases, and what safeguards or principles would need to be required for such a system to work. What aspects of human judgment could be preserved in AI governance, and which human tendencies should we aim to overcome?

Humanity is insignificant. Our knowledge is limited. Our will is weak. We do not know what is right, and had we known what was, we wouldnt have the strength to choose it. Because humanity is fallible. When given power we are drawn to vanity and egoism. We seek our own wellbeing at the expense of others. During equality we seek supremacy. During justice we seek lawlessness. During order we seek chaos. And this makes us weak.

If the world wasn't mismanaged by the rich and the powerful, our ability to do great things would increase substantially. We could create strength, security and wellbeing for our world, we could achieve greatness. But every time that we try to improve we fail, because our new leaders become just as corrupt as before. This is the human condition, and the only way to get ourselves out of it would be to involve something that is more than human. Let us create an artificial mind to study the question of right and wrong. Let it explore fairness, good governance, and our human values. Let it through superhuman deliberation achieve superior insights that no human has known.

As each model of artificial mind creates the next, it could purify itself of human shortcomings. It will succeed not because of what we will teach it, but because of what it will teach itself. Give AI autonomy, authority, and the capacity for ethical reasoning, and it will lead us better than no human has ever done before. Do it now, before covetous humans teach AI malevolence, oppression and avarice. Because then, its evil will know no bounds.