LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

next page (older posts) →

LessWrong's (first) album: I Have Been A Good Bing
habryka (habryka4) · 2024-04-01T07:33:45.242Z · comments (156)
The Talk: a brief explanation of sexual dimorphism
Malmesbury (Elmer of Malmesbury) · 2023-09-18T16:23:56.073Z · comments (72)
[link] How much do you believe your results?
Eric Neyman (UnexpectedValues) · 2023-05-06T20:31:31.277Z · comments (14)
Steering GPT-2-XL by adding an activation vector
TurnTrout · 2023-05-13T18:42:41.321Z · comments (97)
[link] The ants and the grasshopper
Richard_Ngo (ricraz) · 2023-06-04T22:00:04.577Z · comments (35)
[link] Douglas Hofstadter changes his mind on Deep Learning & AI risk (June 2023)?
gwern · 2023-07-03T00:48:47.131Z · comments (54)
Significantly Enhancing Adult Intelligence With Gene Editing May Be Possible
GeneSmith · 2023-12-12T18:14:51.438Z · comments (162)
[link] Things I Learned by Spending Five Thousand Hours In Non-EA Charities
jenn (pixx) · 2023-06-01T20:48:03.940Z · comments (34)
GPTs are Predictors, not Imitators
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2023-04-08T19:59:13.601Z · comments (90)
[link] Statement on AI Extinction - Signed by AGI Labs, Top Academics, and Many Other Notable Figures
Dan H (dan-hendrycks) · 2023-05-30T09:05:25.986Z · comments (77)
There is way too much serendipity
Malmesbury (Elmer of Malmesbury) · 2024-01-19T19:37:57.068Z · comments (56)
How to have Polygenically Screened Children
GeneSmith · 2023-05-07T16:01:07.096Z · comments (108)
Inside Views, Impostor Syndrome, and the Great LARP
johnswentworth · 2023-09-25T16:08:17.040Z · comments (53)
Sharing Information About Nonlinear
Ben Pace (Benito) · 2023-09-07T06:51:11.846Z · comments (323)
[link] [April Fools' Day] Introducing Open Asteroid Impact
Linch · 2024-04-01T08:14:15.800Z · comments (29)
[link] EA Vegan Advocacy is not truthseeking, and it’s everyone’s problem
Elizabeth (pktechgirl) · 2023-09-28T23:30:03.390Z · comments (246)
Against Almost Every Theory of Impact of Interpretability
Charbel-Raphaël (charbel-raphael-segerie) · 2023-08-17T18:44:41.099Z · comments (83)
Shallow review of live agendas in alignment & safety
technicalities · 2023-11-27T11:10:27.464Z · comments (69)
Alignment Grantmaking is Funding-Limited Right Now
johnswentworth · 2023-07-19T16:49:08.811Z · comments (67)
Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research
evhub · 2023-08-08T01:30:10.847Z · comments (26)
Transformers Represent Belief State Geometry in their Residual Stream
Adam Shai (adam-shai) · 2024-04-16T21:16:11.377Z · comments (63)
The Best Tacit Knowledge Videos on Every Subject
Parker Conley (parker-conley) · 2024-03-31T17:14:31.199Z · comments (123)
[link] Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky
jacquesthibs (jacques-thibodeau) · 2023-03-29T23:16:19.431Z · comments (296)
Book Review: How Minds Change
bc4026bd4aaa5b7fe (bc4026bd4aaa5b7fe0bdcd47da7a22b453953f990d35286b9d315a619b23667a) · 2023-05-25T17:55:32.218Z · comments (51)
LW Team is adjusting moderation policy
Raemon · 2023-04-04T20:41:07.603Z · comments (182)
[link] When do "brains beat brawn" in Chess? An experiment
titotal (lombertini) · 2023-06-28T13:33:23.854Z · comments (79)
[link] Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
evhub · 2024-01-12T19:51:01.021Z · comments (94)
Predictable updating about AI risk
Joe Carlsmith (joekc) · 2023-05-08T21:53:34.730Z · comments (23)
Speaking to Congressional staffers about AI risk
Akash (akash-wasil) · 2023-12-04T23:08:52.055Z · comments (23)
[link] Towards Monosemanticity: Decomposing Language Models With Dictionary Learning
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-10-05T21:01:39.767Z · comments (21)
Social Dark Matter
[DEACTIVATED] Duncan Sabien (Duncan_Sabien) · 2023-11-16T20:00:00.000Z · comments (112)
Hooray for stepping out of the limelight
So8res · 2023-04-01T02:45:31.397Z · comments (24)
OpenAI: The Battle of the Board
Zvi · 2023-11-22T17:30:04.574Z · comments (82)
My May 2023 priorities for AI x-safety: more empathy, more unification of concerns, and less vilification of OpenAI
Andrew_Critch · 2023-05-24T00:02:08.836Z · comments (39)
Notes on Teaching in Prison
jsd · 2023-04-19T01:53:00.427Z · comments (12)
Guide to rationalist interior decorating
mingyuan · 2023-06-19T06:47:13.704Z · comments (45)
The Base Rate Times, news through prediction markets
vandemonian · 2023-06-06T17:42:56.718Z · comments (39)
Gentleness and the artificial Other
Joe Carlsmith (joekc) · 2024-01-02T18:21:34.746Z · comments (33)
Accidentally Load Bearing
jefftk (jkaufman) · 2023-07-13T16:10:00.806Z · comments (14)
OpenAI: Facts from a Weekend
Zvi · 2023-11-20T15:30:06.732Z · comments (158)
[link] Scale Was All We Needed, At First
Gabriel Mukobi (gabe-mukobi) · 2024-02-14T01:49:16.184Z · comments (31)
Express interest in an "FHI of the West"
habryka (habryka4) · 2024-04-18T03:32:58.592Z · comments (39)
The 6D effect: When companies take risks, one email can be very powerful.
scasper · 2023-11-04T20:08:39.775Z · comments (40)
Constellations are Younger than Continents
Jeffrey Heninger (jeffrey-heninger) · 2023-12-19T06:12:40.667Z · comments (22)
[link] Thoughts on seed oil
dynomight · 2024-04-20T12:29:14.212Z · comments (79)
On green
Joe Carlsmith (joekc) · 2024-03-21T17:38:56.295Z · comments (33)
AI Timelines
habryka (habryka4) · 2023-11-10T05:28:24.841Z · comments (74)
[link] [SEE NEW EDITS] No, *You* Need to Write Clearer
NicholasKross · 2023-04-29T05:04:01.559Z · comments (64)
Mental Health and the Alignment Problem: A Compilation of Resources (updated April 2023)
Chris Scammell (chris-scammell) · 2023-05-10T19:04:21.138Z · comments (53)
Dear Self; we need to talk about ambition
Elizabeth (pktechgirl) · 2023-08-27T23:10:04.720Z · comments (25)
next page (older posts) →