Posts

Intelligence allocation from a Mean Field Game Theory perspective 2023-06-13T19:52:14.260Z
Aligning alignment with performance 2022-12-14T22:19:35.312Z
Formalizing Alignment 2022-08-10T18:50:02.390Z
Abstraction sacrifices causal clarity 2022-07-31T19:24:01.783Z
Questions for a Theory of Narratives 2022-07-29T19:31:07.472Z

Comments

Comment by Marv K on Intelligence allocation from a Mean Field Game Theory perspective · 2023-06-19T16:45:21.139Z · LW · GW

I agree on both counts. You're right that I should model the alignment of the system as well as its intelligence. I guess the alignment could be thought of as minimizing the distance of high dimensional vectors representing the players' and the AI's values. So each user (and the AI, too) could have a value vector associated with it, and the cost functions of the user could then incorporate how much they care about their own alignment (to the rest of the users), and the cost function of the AI needs to be tuned so that it is enough aligned when it reaches a critical threshold of intelligence. That way, you could express how important it is that the AI is aligned, as a function of its intelligence. 

Comment by Marv K on Generalizing From One Example · 2022-09-14T17:53:00.637Z · LW · GW
Comment by Marv K on Variational Bayesian methods · 2022-08-26T11:18:08.804Z · LW · GW

Nice writeup. I wasn't even aware k-means clustering can be viewed from the Variational Bayes framework. In case more perspectives are useful to any readers: When I first tried to learn about this, I found the Pyro Introduction very helpful; because it is split up over a lot of files, I put together these slides for Bayesian Neural Networks, which also start out with a motivation for Variational Bayes.

Comment by Marv K on Announcing the Alignment of Complex Systems Research Group · 2022-08-10T18:55:10.309Z · LW · GW

I've been thinking about alignment of subsystems in a very similar style and am really excited to see someone  else thinking along this way. I started a comment with my own thoughts on this approach; but it got out of hand quickly; so I made a separate post: https://www.lesswrong.com/posts/AZfq4jLjqsrt5fjGz/formalizing-alignment 

Would be keen on having any sort of feedback.

Comment by Marv K on Questions for a Theory of Narratives · 2022-07-31T11:42:27.705Z · LW · GW

Thanks for the pointers! The overviews in both sources are great. I especially like Rumelhart's Story Grammar. Though from what I gather from Mark Riedl's post is that the field is mostly about structure/grammar inherent to stories as objects that exist pretty much in a vacuum, and does not explicitly focus on making connections to some sort of models of agents that communicate using these stories.