LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Fucking Goddamn Basics of Rationalist Discourse
LoganStrohl (BrienneYudkowsky) · 2023-02-04T01:47:32.578Z · comments (103)
Sharing Information About Nonlinear
Ben Pace (Benito) · 2023-09-07T06:51:11.846Z · comments (323)
Feature Selection
Zack_M_Davis · 2021-11-01T00:22:29.993Z · comments (24)
You don't know how bad most things are nor precisely how they're bad.
Solenoid_Entity · 2024-08-04T14:12:54.136Z · comments (48)
[link] EA Vegan Advocacy is not truthseeking, and it’s everyone’s problem
Elizabeth (pktechgirl) · 2023-09-28T23:30:03.390Z · comments (250)
Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research
evhub · 2023-08-08T01:30:10.847Z · comments (29)
“PR” is corrosive; “reputation” is not.
AnnaSalamon · 2021-02-14T03:32:24.985Z · comments (95)
[link] When do "brains beat brawn" in Chess? An experiment
titotal (lombertini) · 2023-06-28T13:33:23.854Z · comments (103)
[link] I got dysentery so you don’t have to
eukaryote · 2024-10-22T04:55:58.422Z · comments (4)
What Goes Without Saying
sarahconstantin · 2024-12-20T18:00:06.363Z · comments (27)
The Case Against AI Control Research
johnswentworth · 2025-01-21T16:03:10.143Z · comments (74)
Models Don't "Get Reward"
Sam Ringer · 2022-12-30T10:37:11.798Z · comments (61)
Alignment Grantmaking is Funding-Limited Right Now
johnswentworth · 2023-07-19T16:49:08.811Z · comments (68)
Book Review: How Minds Change
bc4026bd4aaa5b7fe (bc4026bd4aaa5b7fe0bdcd47da7a22b453953f990d35286b9d315a619b23667a) · 2023-05-25T17:55:32.218Z · comments (52)
Six Dimensions of Operational Adequacy in AGI Projects
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-05-30T17:00:30.833Z · comments (66)
The Parable of the King and the Random Process
moridinamael · 2023-03-01T22:18:59.734Z · comments (26)
Epistemic Legibility
Elizabeth (pktechgirl) · 2022-02-09T18:10:06.591Z · comments (30)
[link] Industrial literacy
jasoncrawford · 2020-09-30T16:39:06.520Z · comments (130)
Guide to rationalist interior decorating
mingyuan · 2023-06-19T06:47:13.704Z · comments (49)
Universal Basic Income and Poverty
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2024-07-26T07:23:50.151Z · comments (136)
Leaky Delegation: You are not a Commodity
Darmani · 2021-01-25T02:04:55.942Z · comments (38)
[link] Great minds might not think alike
Eric Neyman (UnexpectedValues) · 2020-12-26T19:51:05.978Z · comments (45)
On not getting contaminated by the wrong obesity ideas
Natália (Natália Mendonça) · 2023-01-28T20:18:21.322Z · comments (69)
On how various plans miss the hard bits of the alignment challenge
So8res · 2022-07-12T02:49:50.454Z · comments (89)
[link] Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
evhub · 2024-01-12T19:51:01.021Z · comments (95)
Would catching your AIs trying to escape convince AI developers to slow down or undeploy?
Buck · 2024-08-26T16:46:18.872Z · comments (77)
LW Team is adjusting moderation policy
Raemon · 2023-04-04T20:41:07.603Z · comments (185)
Why Agent Foundations? An Overly Abstract Explanation
johnswentworth · 2022-03-25T23:17:10.324Z · comments (56)
Speaking to Congressional staffers about AI risk
[deleted] · 2023-12-04T23:08:52.055Z · comments (25)
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger (RobbBB) · 2022-12-01T23:11:44.279Z · comments (33)
An Unexpected Victory: Container Stacking at the Port of Long Beach
Zvi · 2021-10-28T14:40:00.497Z · comments (41)
Heads I Win, Tails?—Never Heard of Her; Or, Selective Reporting and the Tragedy of the Green Rationalists
Zack_M_Davis · 2019-09-24T04:12:07.560Z · comments (40)
Lies, Damn Lies, and Fabricated Options
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2021-10-17T02:47:24.909Z · comments (132)
The Field of AI Alignment: A Postmortem, and What To Do About It
johnswentworth · 2024-12-26T18:48:07.614Z · comments (158)
EfficientZero: How It Works
1a3orn · 2021-11-26T15:17:08.321Z · comments (50)
Science in a High-Dimensional World
johnswentworth · 2021-01-08T17:52:02.261Z · comments (53)
LessWrong is providing feedback and proofreading on drafts as a service
Ruby · 2021-09-07T01:33:10.666Z · comments (53)
Gentleness and the artificial Other
Joe Carlsmith (joekc) · 2024-01-02T18:21:34.746Z · comments (33)
Two-year update on my personal AI timelines
Ajeya Cotra (ajeya-cotra) · 2022-08-02T23:07:48.698Z · comments (60)
[link] Is Success the Enemy of Freedom? (Full)
alkjash · 2020-10-26T20:25:50.503Z · comments (69)
AI Timelines
habryka (habryka4) · 2023-11-10T05:28:24.841Z · comments (133)
Predictable updating about AI risk
Joe Carlsmith (joekc) · 2023-05-08T21:53:34.730Z · comments (25)
Study Guide
johnswentworth · 2021-11-06T01:23:09.552Z · comments (48)
[link] Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky
jacquesthibs (jacques-thibodeau) · 2023-03-29T23:16:19.431Z · comments (297)
Politics is way too meta
Rob Bensinger (RobbBB) · 2021-03-17T07:04:42.187Z · comments (46)
[link] Intentionally Making Close Friends
Neel Nanda (neel-nanda-1) · 2021-06-27T23:06:49.269Z · comments (35)
[link] Towards Monosemanticity: Decomposing Language Models With Dictionary Learning
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-10-05T21:01:39.767Z · comments (22)
Non-Disparagement Canaries for OpenAI
aysja · 2024-05-30T19:20:13.022Z · comments (51)
[link] Scale Was All We Needed, At First
Gabe M (gabe-mukobi) · 2024-02-14T01:49:16.184Z · comments (33)
Accidentally Load Bearing
jefftk (jkaufman) · 2023-07-13T16:10:00.806Z · comments (17)
← previous page (newer posts) · next page (older posts) →