LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

What's Hard About The Shutdown Problem
johnswentworth · 2023-10-20T21:13:27.624Z · comments (33)
Hedonic asymmetries
paulfchristiano · 2020-01-26T02:10:01.323Z · comments (22)
On attunement
Joe Carlsmith (joekc) · 2024-03-25T12:47:34.856Z · comments (8)
The Social Alignment Problem
irving (judith) · 2023-04-28T14:16:17.825Z · comments (13)
Rant on Problem Factorization for Alignment
johnswentworth · 2022-08-05T19:23:24.262Z · comments (53)
Naive Hypotheses on AI Alignment
Shoshannah Tekofsky (DarkSym) · 2022-07-02T19:03:49.458Z · comments (29)
Recommending Understand, a Game about Discerning the Rules
MondSemmel · 2021-10-28T14:53:16.901Z · comments (55)
Why Artists Study Anatomy
Sisi Cheng (sisi-cheng) · 2020-05-18T18:44:23.109Z · comments (10)
[link] "Deep Learning" Is Function Approximation
Zack_M_Davis · 2024-03-21T17:50:36.254Z · comments (28)
Reframing Impact
TurnTrout · 2019-09-20T19:03:27.898Z · comments (15)
Connectomics seems great from an AI x-risk perspective
Steven Byrnes (steve2152) · 2023-04-30T14:38:39.738Z · comments (7)
[link] My cover story in Jacobin on AI capitalism and the x-risk debates
garrison · 2024-02-12T23:34:16.526Z · comments (5)
Salvage Epistemology
jimrandomh · 2022-04-30T02:10:41.996Z · comments (119)
Marriage, the Giving What We Can Pledge, and the damage caused by vague public commitments
Jeffrey Ladish (jeff-ladish) · 2022-07-11T19:38:42.468Z · comments (27)
Clem's Memo
abstractapplic · 2022-04-16T11:59:55.704Z · comments (8)
The nihilism of NeurIPS
charlieoneill (kingchucky211) · 2024-12-20T23:58:11.858Z · comments (7)
Bigger Livers?
sarahconstantin · 2024-11-08T21:50:09.814Z · comments (13)
Total horse takeover
KatjaGrace · 2019-11-05T00:10:01.319Z · comments (14)
[link] [Link] Why I’m optimistic about OpenAI’s alignment approach
janleike · 2022-12-05T22:51:15.769Z · comments (15)
Trying to Keep the Garden Well
Tobias H (clearthis) · 2022-01-16T05:42:11.851Z · comments (5)
[Valence series] 1. Introduction
Steven Byrnes (steve2152) · 2023-12-04T15:40:21.274Z · comments (16)
Geoff Hinton Quits Google
Adam Shai (adam-shai) · 2023-05-01T21:03:47.806Z · comments (14)
Truth and Advantage: Response to a draft of "AI safety seems hard to measure"
So8res · 2023-03-22T03:36:02.945Z · comments (10)
MIRI’s 2024 End-of-Year Update
Rob Bensinger (RobbBB) · 2024-12-03T04:33:47.499Z · comments (2)
Announcing the London Initiative for Safe AI (LISA)
James Fox · 2024-02-02T23:17:47.011Z · comments (0)
[Intro to brain-like-AGI safety] 15. Conclusion: Open problems, how to help, AMA
Steven Byrnes (steve2152) · 2022-05-17T15:11:12.397Z · comments (10)
Explaining a Math Magic Trick
Robert_AIZI · 2024-05-05T19:41:52.048Z · comments (10)
List of resolved confusions about IDA
Wei Dai (Wei_Dai) · 2019-09-30T20:03:10.506Z · comments (18)
Would we even want AI to solve all our problems?
So8res · 2023-04-21T18:04:11.636Z · comments (15)
We have some evidence that masks work
technicalities · 2021-07-11T18:36:46.942Z · comments (13)
Contrast Pairs Drive the Empirical Performance of Contrast Consistent Search (CCS)
Scott Emmons · 2023-05-31T17:09:02.288Z · comments (1)
[link] CIV: a story
Richard_Ngo (ricraz) · 2024-06-15T22:36:50.415Z · comments (6)
How To Make Prediction Markets Useful For Alignment Work
johnswentworth · 2022-10-18T19:01:01.292Z · comments (18)
A summary of every "Highlights from the Sequences" post
Akash (akash-wasil) · 2022-07-15T23:01:04.392Z · comments (7)
Desperation hamster wheels
Nicole Ross (nicole-ross) · 2020-10-28T02:32:27.000Z · comments (5)
April 15, 2040
Nisan · 2021-05-04T21:18:08.912Z · comments (25)
[link] Seven lessons I didn't learn from election day
Eric Neyman (UnexpectedValues) · 2024-11-14T18:39:07.053Z · comments (33)
I don’t find the lie detection results that surprising (by an author of the paper)
JanB (JanBrauner) · 2023-10-04T17:10:51.262Z · comments (8)
Meta-level adversarial evaluation of oversight techniques might allow robust measurement of their adequacy
Buck · 2023-07-26T17:02:56.456Z · comments (19)
Clarifying “What failure looks like”
Sam Clarke · 2020-09-20T20:40:48.295Z · comments (14)
Tessellating Hills: a toy model for demons in imperfect search
DaemonicSigil · 2020-02-20T00:12:50.125Z · comments (18)
Thinking About Filtered Evidence Is (Very!) Hard
abramdemski · 2020-03-19T23:20:05.562Z · comments (32)
[link] Announcing Epoch: A research organization investigating the road to Transformative AI
Jsevillamol · 2022-06-27T13:55:51.451Z · comments (2)
Good Heart Week: Extending the Experiment
Ben Pace (Benito) · 2022-04-02T07:13:48.353Z · comments (92)
Given the Restrict Act, Don’t Ban TikTok
Zvi · 2023-04-04T14:40:03.162Z · comments (9)
[link] [Linkpost] A survey on over 300 works about interpretability in deep networks
scasper · 2022-09-12T19:07:09.156Z · comments (7)
Access to powerful AI might make computer security radically easier
Buck · 2024-06-08T06:00:19.310Z · comments (14)
Paper: Teaching GPT3 to express uncertainty in words
Owain_Evans · 2022-05-31T13:27:17.191Z · comments (7)
Comments on Anthropic's Scaling Monosemanticity
Robert_AIZI · 2024-06-03T12:15:44.708Z · comments (8)
Basic facts about language models during training
beren · 2023-02-21T11:46:12.256Z · comments (15)
← previous page (newer posts) · next page (older posts) →