Posts

RerM's Shortform 2025-04-28T03:30:44.100Z

Comments

Comment by RerM (robert-m) on RerM's Shortform · 2025-04-28T03:30:44.099Z · LW · GW

Generally, hypothetical hostile AGI is assumed to be made on software/hardware that's more advanced from what we have now. This makes sense, as Chat-GPT is very stupid in a lot of ways.
Has anyone considered purposefully creating a hostile AGI on this "stupid" software so we can wargame how a highly advanced, hostile AGI would act? Obviously the difference between what we have now and what we may have later will be quite large, but I think we could create a project were we "fight" stupid AIs, then slowly move up the intelligence ladder as new models come out, using our newfound knowledge of fighting hostile intelligence to mitigate the risk that comes with creating hostile AIs.

Has anyone ever thought of this? Also, what are your thoughts on this? Alignment and AI are not my specialties, but I thought this idea sounded interesting enough to share.

Comment by RerM (robert-m) on That Alien Message - The Animation · 2024-09-14T14:20:31.696Z · LW · GW

Very interesting. Question: How does putting humans into cryonic suspension relate or contribute to the metaphor, if at all?