LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Future ML Systems Will Be Qualitatively Different
jsteinhardt · 2022-01-11T19:50:11.377Z · comments (10)
AI #14: A Very Good Sentence
Zvi · 2023-06-01T21:30:04.548Z · comments (30)
Going Crazy and Getting Better Again
Evenstar · 2023-07-02T18:55:25.790Z · comments (10)
Why was the AI Alignment community so unprepared for this moment?
Ras1513 · 2023-07-15T00:26:29.769Z · comments (64)
Theses on Sleep
guzey · 2022-02-11T12:58:15.300Z · comments (104)
AGI and the EMH: markets are not expecting aligned or unaligned AI in the next 30 years
basil.halperin (bhalperin) · 2023-01-10T16:06:52.329Z · comments (44)
[question] How do we prepare for final crunch time?
Eli Tyre (elityre) · 2021-03-30T05:47:54.654Z · answers+comments (30)
Mental health benefits and downsides of psychedelic use in ACX readers: survey results
RationalElf · 2021-10-25T22:55:09.522Z · comments (18)
Mnestics
Jarred Filmer (4thWayWastrel) · 2022-10-23T00:30:11.159Z · comments (5)
But why would the AI kill us?
So8res · 2023-04-17T18:42:39.720Z · comments (86)
[question] How Hard Would It Be To Make A COVID Vaccine For Oneself?
johnswentworth · 2020-12-21T16:19:10.415Z · answers+comments (29)
Cup-Stacking Skills (or, Reflexive Involuntary Mental Motions)
[DEACTIVATED] Duncan Sabien (Duncan_Sabien) · 2021-10-11T07:16:45.950Z · comments (36)
Honoring Petrov Day on LessWrong, in 2020
Ben Pace (Benito) · 2020-09-26T08:01:36.838Z · comments (100)
AI Safety "Success Stories"
Wei Dai (Wei_Dai) · 2019-09-07T02:54:15.003Z · comments (27)
[link] Philosophy of Therapy
DaystarEld · 2020-10-10T20:12:38.204Z · comments (27)
Situating LessWrong in contemporary philosophy: An interview with Jon Livengood
Suspended Reason (suspended-reason) · 2020-07-01T00:37:00.695Z · comments (21)
We have achieved Noob Gains in AI
phdead · 2022-05-18T20:56:49.143Z · comments (20)
My Effortless Weightloss Story: A Quick Runthrough
CuoreDiVetro · 2023-09-30T23:02:45.128Z · comments (76)
Geometric Exploration, Arithmetic Exploitation
Scott Garrabrant · 2022-11-24T15:36:30.334Z · comments (4)
[link] Scott Aaronson is joining OpenAI to work on AI safety
peterbarnett · 2022-06-18T04:06:55.465Z · comments (31)
Soft optimization makes the value target bigger
Jeremy Gillen (jeremy-gillen) · 2023-01-02T16:06:50.229Z · comments (20)
unRLHF - Efficiently undoing LLM safeguards
Pranav Gade (pranav-gade) · 2023-10-12T19:58:08.811Z · comments (15)
Full-time AGI Safety!
Steven Byrnes (steve2152) · 2021-03-01T12:42:14.813Z · comments (3)
Why I take short timelines seriously
NicholasKees (nick_kees) · 2024-01-28T22:27:21.098Z · comments (29)
Revealing Intentionality In Language Models Through AdaVAE Guided Sampling
jdp · 2023-10-20T07:32:28.749Z · comments (14)
GPT-3 Catching Fish in Morse Code
Megan Kinniment (megan-kinniment) · 2022-06-30T21:22:49.054Z · comments (27)
AMA: Paul Christiano, alignment researcher
paulfchristiano · 2021-04-28T18:55:39.707Z · comments (197)
Announcing $5,000 bounty for (responsibly) ending malaria
lc · 2022-09-24T04:28:22.189Z · comments (40)
Reducing sycophancy and improving honesty via activation steering
Nina Rimsky (NinaR) · 2023-07-28T02:46:23.122Z · comments (14)
Book review: "A Thousand Brains" by Jeff Hawkins
Steven Byrnes (steve2152) · 2021-03-04T05:10:44.929Z · comments (18)
[link] When discussing AI risks, talk about capabilities, not intelligence
Vika · 2023-08-11T13:38:48.844Z · comments (7)
[link] Manifold: If okay AGI, why?
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2023-03-25T22:43:53.820Z · comments (37)
Parable of the Dammed
johnswentworth · 2020-12-10T00:08:44.493Z · comments (29)
[link] Popular education in Sweden: much more than you wanted to know
Henrik Karlsson (henrik-karlsson) · 2022-05-17T20:07:50.318Z · comments (3)
Experiences raising children in shared housing
juliawise · 2021-12-21T17:09:05.008Z · comments (4)
Covid 6/25: The Dam Breaks
Zvi · 2020-06-25T18:30:02.899Z · comments (21)
Goodhart's Law inside the human mind
Kaj_Sotala · 2023-04-17T13:48:13.183Z · comments (13)
ARC tests to see if GPT-4 can escape human control; GPT-4 failed to do so
Christopher King (christopher-king) · 2023-03-15T00:29:23.523Z · comments (22)
Unifying Bargaining Notions (2/2)
Diffractor · 2022-07-27T03:40:30.524Z · comments (19)
[question] Why The Focus on Expected Utility Maximisers?
DragonGod · 2022-12-27T15:49:36.536Z · answers+comments (84)
“Reframing Superintelligence” + LLMs + 4 years
Eric Drexler · 2023-07-10T13:42:09.739Z · comments (8)
On Investigating Conspiracy Theories
Zvi · 2023-02-20T12:50:00.891Z · comments (38)
[link] Why Weren't Hot Air Balloons Invented Sooner?
Lost Futures (aeviternity1) · 2022-10-18T00:41:43.156Z · comments (52)
The LessWrong 2022 Review
habryka (habryka4) · 2023-12-05T04:00:00.000Z · comments (43)
[link] On hiding the source of knowledge
jessicata (jessica.liu.taylor) · 2020-01-26T02:48:51.310Z · comments (40)
Transcript of Sam Altman's interview touching on AI safety
Andy_McKenzie · 2023-01-20T16:14:18.974Z · comments (41)
I Would Have Solved Alignment, But I Was Worried That Would Advance Timelines
307th · 2023-10-20T16:37:46.541Z · comments (32)
Can you get AGI from a Transformer?
Steven Byrnes (steve2152) · 2020-07-23T15:27:51.712Z · comments (40)
The case against AI alignment
andrew sauer (andrew-sauer) · 2022-12-24T06:57:53.405Z · comments (110)
[link] In Defense of Attempting Hard Things, and my story of the Leverage ecosystem
Cathleen · 2021-12-17T23:08:20.150Z · comments (43)
← previous page (newer posts) · next page (older posts) →