Posts

Riga EA/ACX Social August 2024 2024-08-04T12:01:38.652Z
Riga EA/ACX Meetup July 2024 2024-07-16T13:21:11.949Z
The Future of Work: How Can Policymakers Prepare for AI's Impact on Labor Markets? 2024-06-24T14:18:55.023Z
Riga EA/ACX Meetup June 2024 2024-06-20T16:33:09.773Z
Riga, Latvia – ACX Meetups Everywhere 2023 2023-08-23T20:07:20.967Z

Comments

Comment by Arturs (arturs-kanepajs) on Access to powerful AI might make computer security radically easier · 2024-06-09T07:55:03.362Z · LW · GW

These frontier models could still be vulnerable to stealth (e.g. “sleeper agent”) attacks, specialist models, and stealth attacks by specialist models. The balance depends on the ability gap – if the top model is way ahead of others, then maybe defence dominates attack efforts. But a big ability gap does not seem to be playing out, instead there are several frontier models near-frontier, and lots of (more or less) open source stuff not far behind.

Comment by Arturs (arturs-kanepajs) on Predictable updating about AI risk · 2023-05-19T11:42:38.298Z · LW · GW

Seems like a stark case of contrast between Bayesianism and the way a frequentist might approach things. I.e. do not reject the null hypothesis of no significant probability until convinced by evidence, either formal arguments or by seeing real-life mishaps. Labeling something as having P(x)~0 probably helps to compartmentalize things, focus to other tasks at hand. But can lead to huge risks being neglected, like in this case of AI Alignment.

Edit: "premortem" seems like a useful exercise to align mind & gut