LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

next page (older posts) →

AGI Ruin: A List of Lethalities
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-06-05T22:05:52.224Z · comments (690)
Where I agree and disagree with Eliezer
paulfchristiano · 2022-06-19T19:15:55.698Z · comments (219)
It’s Probably Not Lithium
Natália (Natália Mendonça) · 2022-06-28T21:24:10.246Z · comments (186)
Security Mindset: Lessons from 20+ years of Software Security Failures Relevant to AGI Alignment
elspood · 2022-06-21T23:55:39.918Z · comments (42)
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res · 2022-06-15T13:10:18.658Z · comments (53)
What Are You Tracking In Your Head?
johnswentworth · 2022-06-28T19:30:06.164Z · comments (81)
Humans are very reliable agents
alyssavance · 2022-06-16T22:02:10.892Z · comments (35)
Comment reply: my low-quality thoughts on why CFAR didn't get farther with a "real/efficacious art of rationality"
AnnaSalamon · 2022-06-09T02:12:35.151Z · comments (62)
Slow motion videos as AI risk intuition pumps
Andrew_Critch · 2022-06-14T19:31:13.616Z · comments (41)
Contra Hofstadter on GPT-3 Nonsense
rictic · 2022-06-15T21:53:30.646Z · comments (24)
AGI Safety FAQ / all-dumb-questions-allowed thread
Aryeh Englander (alenglander) · 2022-06-07T05:47:13.350Z · comments (526)
The inordinately slow spread of good AGI conversations in ML
Rob Bensinger (RobbBB) · 2022-06-21T16:09:57.859Z · comments (62)
AI Could Defeat All Of Us Combined
HoldenKarnofsky · 2022-06-09T15:50:12.952Z · comments (42)
Announcing the Inverse Scaling Prize ($250k Prize Pool)
Ethan Perez (ethan-perez) · 2022-06-27T15:58:19.135Z · comments (14)
The prototypical catastrophic AI action is getting root access to its datacenter
Buck · 2022-06-02T23:46:31.360Z · comments (13)
A transparency and interpretability tech tree
evhub · 2022-06-16T23:44:14.961Z · comments (11)
On A List of Lethalities
Zvi · 2022-06-13T12:30:01.624Z · comments (49)
Why all the fuss about recursive self-improvement?
So8res · 2022-06-12T20:53:42.392Z · comments (62)
Nonprofit Boards are Weird
HoldenKarnofsky · 2022-06-23T14:40:11.593Z · comments (26)
Staying Split: Sabatini and Social Justice
[DEACTIVATED] Duncan Sabien (Duncan_Sabien) · 2022-06-08T08:32:58.633Z · comments (28)
LessWrong Has Agree/Disagree Voting On All New Comment Threads
Ben Pace (Benito) · 2022-06-24T00:43:17.136Z · comments (217)
Godzilla Strategies
johnswentworth · 2022-06-11T15:44:16.385Z · comments (71)
[question] why assume AGIs will optimize for fixed goals?
nostalgebraist · 2022-06-10T01:28:10.961Z · answers+comments (55)
Public beliefs vs. Private beliefs
Eli Tyre (elityre) · 2022-06-01T21:33:32.661Z · comments (30)
Deep Learning Systems Are Not Less Interpretable Than Logic/Probability/Etc
johnswentworth · 2022-06-04T05:41:56.713Z · comments (53)
A descriptive, not prescriptive, overview of current AI Alignment Research
Jan (jan-2) · 2022-06-06T21:59:22.344Z · comments (21)
Limits to Legibility
Jan_Kulveit · 2022-06-29T17:42:19.338Z · comments (11)
Announcing the LessWrong Curated Podcast
Ben Pace (Benito) · 2022-06-22T22:16:58.170Z · comments (26)
[link] AI-Written Critiques Help Humans Notice Flaws
paulfchristiano · 2022-06-25T17:22:56.959Z · comments (5)
Contra EY: Can AGI destroy us without trial & error?
Nikita Sokolsky (nikita-sokolsky) · 2022-06-13T18:26:09.460Z · comments (72)
Steam
abramdemski · 2022-06-20T17:38:58.548Z · comments (13)
Confused why a "capabilities research is good for alignment progress" position isn't discussed more
Kaj_Sotala · 2022-06-02T21:41:44.784Z · comments (27)
Intergenerational trauma impeding cooperative existential safety efforts
Andrew_Critch · 2022-06-03T08:13:25.439Z · comments (29)
"Pivotal Acts" means something specific
Raemon · 2022-06-07T21:56:00.574Z · comments (23)
Let's See You Write That Corrigibility Tag
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-06-19T21:11:03.505Z · comments (69)
Will Capabilities Generalise More?
Ramana Kumar (ramana-kumar) · 2022-06-29T17:12:56.255Z · comments (39)
Conversation with Eliezer: What do you want the system to do?
Akash (akash-wasil) · 2022-06-25T17:36:14.145Z · comments (38)
[link] Scott Aaronson is joining OpenAI to work on AI safety
peterbarnett · 2022-06-18T04:06:55.465Z · comments (31)
Leaving Google, Joining the Nucleic Acid Observatory
jefftk (jkaufman) · 2022-06-10T17:00:01.893Z · comments (4)
Who models the models that model models? An exploration of GPT-3's in-context model fitting ability
Lovre · 2022-06-07T19:37:49.276Z · comments (16)
CFAR Handbook: Introduction
CFAR!Duncan (CFAR 2017) · 2022-06-28T16:53:53.312Z · comments (12)
wrapper-minds are the enemy
nostalgebraist · 2022-06-17T01:58:04.919Z · comments (41)
Yes, AI research will be substantially curtailed if a lab causes a major disaster
lc · 2022-06-14T22:17:01.273Z · comments (31)
Relationship Advice Repository
Ruby · 2022-06-20T14:39:36.548Z · comments (36)
[link] Announcing Epoch: A research organization investigating the road to Transformative AI
Jsevillamol · 2022-06-27T13:55:51.451Z · comments (2)
Pivotal outcomes and pivotal processes
Andrew_Critch · 2022-06-17T23:43:19.230Z · comments (31)
Units of Exchange
CFAR!Duncan (CFAR 2017) · 2022-06-28T16:53:53.069Z · comments (28)
Contest: An Alien Message
DaemonicSigil · 2022-06-27T05:54:54.144Z · comments (100)
The Mountain Troll
lsusr · 2022-06-11T09:14:01.479Z · comments (25)
Against Active Shooter Drills
Zvi · 2022-06-16T13:40:02.690Z · comments (30)
next page (older posts) →