What if we could use the theory of Mechanism Design from Game Theory as a medium achieve AI Alignment?

post by farari7 · 2021-04-04T12:56:05.263Z · LW · GW · No comments

This is a question post.

Contents

No comments

I watch some nice conferences from the MIRI team about the challenges in the area of AI Alignment (specifying the right utility function, 

making an AI reliable for the task, ways to teach an AI etcetera). While looking to these information, and given the reward and utility functions 

paradigm they used,  it came to my mind, what if we could use some of the theory developed in Mechanism Design to incentive desirable Social Choice Functions for AI agents?

and guarantee nice properties like knowing the agent won't lie about private information (maybe intentions, source code, etc.) or other axiomatic principles like Pareto, Non Dictatorial etc.

I am wondering if anyone has given a though to this and if they are willing to share what their analysis is?

 (for anyone interested in this area I leave a link with a good introduction).

Answers

No comments

Comments sorted by top scores.