Roman Malov's Shortform
post by Roman Malov · 2024-12-19T21:14:54.805Z · LW · GW · 1 commentsContents
1 comment
1 comments
Comments sorted by top scores.
comment by Roman Malov · 2024-12-19T21:14:55.985Z · LW(p) · GW(p)
I recently prepared an overview lecture about research directions in AI alignment for the Moscow AI Safety Hub. I had limited time, so I did the following: I reviewed all the sites on the AI safety map, examined the 'research' sections, and attempted to classify the problems they tackle and the research paths they pursue. I encountered difficulties in this process, partly because most sites lack a brief summary of their activities and objectives (Conjecture is one of the counterexamples). I believe that the field of AI safety would greatly benefit from improved communication, and providing a brief summary of a research direction seems like low-hanging fruit.