Looking for judges for critiques of Alignment Plans
post by Iknownothing · 2023-08-17T22:35:40.666Z · LW · GW · 0 commentsContents
No comments
Hello!
AI-Plans.com recently held a "critique-a-thon," where participants submitted and refined 40+ critiques for AI Alignment plans. Here are the finalized critiques from the event: https://docs.google.com/document/d/1mW4SAxFN_aI6KyYXpl9qz5B9nVdeV9Xyc69GTNme5cA/edit?usp=sharing
We are looking for anyone who might be interested in helping to judge these final 11 critiques.
So far, we have gratefully had the assistance of Dr Peter S Park (MIT postdoc at the Tegmark lab, Harvard PhD) and Aishwarya G (AI Existential Safety Community Member at Future of Life Institute and Governance Course Facilitator at BlueDot Impact AI Safety Fundamentals), as well as some independent alignment researchers.
I would love to hear your thoughts!
Kabir Kumar (Founder, AI-Plans.com)
0 comments
Comments sorted by top scores.