LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

next page (older posts) →

AGI Ruin: A List of Lethalities
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-06-05T22:05:52.224Z · comments (701)
Where I agree and disagree with Eliezer
paulfchristiano · 2022-06-19T19:15:55.698Z · comments (220)
What an actually pessimistic containment strategy looks like
lc · 2022-04-05T00:19:50.212Z · comments (138)
[link] Simulators
janus · 2022-09-02T12:45:33.723Z · comments (162)
Let’s think about slowing down AI
KatjaGrace · 2022-12-22T17:40:04.787Z · comments (182)
The Redaction Machine
Ben (ben-lang) · 2022-09-20T22:03:15.309Z · comments (48)
[link] Luck based medicine: my resentful story of becoming a medical miracle
Elizabeth (pktechgirl) · 2022-10-16T17:40:03.702Z · comments (121)
Losing the root for the tree
Adam Zerner (adamzerner) · 2022-09-20T04:53:53.435Z · comments (31)
Counter-theses on Sleep
Natália (Natália Mendonça) · 2022-03-21T23:21:07.943Z · comments (131)
It’s Probably Not Lithium
Natália (Natália Mendonça) · 2022-06-28T21:24:10.246Z · comments (187)
chinchilla's wild implications
nostalgebraist · 2022-07-31T01:18:28.254Z · comments (128)
(My understanding of) What Everyone in Technical Alignment is Doing and Why
Thomas Larsen (thomas-larsen) · 2022-08-29T01:23:58.073Z · comments (90)
[link] It Looks Like You're Trying To Take Over The World
gwern · 2022-03-09T16:35:35.326Z · comments (120)
DeepMind alignment team opinions on AGI ruin arguments
Vika · 2022-08-12T21:06:40.582Z · comments (37)
[link] Reflections on six months of fatherhood
jasoncrawford · 2022-01-31T05:28:09.154Z · comments (24)
Reward is not the optimization target
TurnTrout · 2022-07-25T00:03:18.307Z · comments (123)
Lies Told To Children
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-04-14T11:25:10.282Z · comments (94)
You Are Not Measuring What You Think You Are Measuring
johnswentworth · 2022-09-20T20:04:22.899Z · comments (44)
[link] A Mechanistic Interpretability Analysis of Grokking
Neel Nanda (neel-nanda-1) · 2022-08-15T02:41:36.245Z · comments (47)
Counterarguments to the basic AI x-risk case
KatjaGrace · 2022-10-14T13:00:05.903Z · comments (124)
Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover
Ajeya Cotra (ajeya-cotra) · 2022-07-18T19:06:14.670Z · comments (94)
Accounting For College Costs
johnswentworth · 2022-04-01T17:28:19.409Z · comments (41)
Security Mindset: Lessons from 20+ years of Software Security Failures Relevant to AGI Alignment
elspood · 2022-06-21T23:55:39.918Z · comments (42)
What DALL-E 2 can and cannot do
Swimmer963 (Miranda Dixon-Luinenburg) (Swimmer963) · 2022-05-01T23:51:22.310Z · comments (303)
Staring into the abyss as a core life skill
benkuhn · 2022-12-22T15:30:05.093Z · comments (21)
MIRI announces new "Death With Dignity" strategy
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-04-02T00:43:19.814Z · comments (545)
What should you change in response to an "emergency"? And AI risk
AnnaSalamon · 2022-07-18T01:11:14.667Z · comments (60)
Why I think strong general AI is coming soon
porby · 2022-09-28T05:40:38.395Z · comments (141)
Looking back on my alignment PhD
TurnTrout · 2022-07-01T03:19:59.497Z · comments (64)
Beware boasting about non-existent forecasting track records
Jotto999 · 2022-05-20T19:20:03.854Z · comments (112)
Optimality is the tiger, and agents are its teeth
Veedrac · 2022-04-02T00:46:27.138Z · comments (42)
Models Don't "Get Reward"
Sam Ringer · 2022-12-30T10:37:11.798Z · comments (61)
Six Dimensions of Operational Adequacy in AGI Projects
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-05-30T17:00:30.833Z · comments (66)
Epistemic Legibility
Elizabeth (pktechgirl) · 2022-02-09T18:10:06.591Z · comments (30)
On how various plans miss the hard bits of the alignment challenge
So8res · 2022-07-12T02:49:50.454Z · comments (88)
Why Agent Foundations? An Overly Abstract Explanation
johnswentworth · 2022-03-25T23:17:10.324Z · comments (56)
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger (RobbBB) · 2022-12-01T23:11:44.279Z · comments (33)
Two-year update on my personal AI timelines
Ajeya Cotra (ajeya-cotra) · 2022-08-02T23:07:48.698Z · comments (60)
Mysteries of mode collapse
janus · 2022-11-08T10:37:57.760Z · comments (57)
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res · 2022-06-15T13:10:18.658Z · comments (54)
We Choose To Align AI
johnswentworth · 2022-01-01T20:06:23.307Z · comments (16)
Don't die with dignity; instead play to your outs
Jeffrey Ladish (jeff-ladish) · 2022-04-06T07:53:05.172Z · comments (60)
What Are You Tracking In Your Head?
johnswentworth · 2022-06-28T19:30:06.164Z · comments (83)
Is AI Progress Impossible To Predict?
alyssavance · 2022-05-15T18:30:12.103Z · comments (39)
Sazen
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2022-12-21T07:54:51.415Z · comments (83)
Toni Kurz and the Insanity of Climbing Mountains
GeneSmith · 2022-07-03T20:51:58.429Z · comments (67)
Humans are very reliable agents
alyssavance · 2022-06-16T22:02:10.892Z · comments (35)
12 interesting things I learned studying the discovery of nature's laws
Ben Pace (Benito) · 2022-02-19T23:39:47.841Z · comments (40)
Changing the world through slack & hobbies
Steven Byrnes (steve2152) · 2022-07-21T18:11:05.636Z · comments (13)
Safetywashing
Adam Scholl (adam_scholl) · 2022-07-01T11:56:33.495Z · comments (20)
next page (older posts) →