LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Simulacra Levels Summary
Zvi · 2023-01-30T13:40:00.774Z · comments (12)
Compounding Resource X
Raemon · 2023-01-11T03:14:08.565Z · comments (5)
[link] Opportunity Cost Blackmail
adamShimi · 2023-01-02T13:48:51.811Z · comments (11)
A general comment on discussions of genetic group differences
anonymous8101 (petrov-amethyst) · 2023-01-14T02:11:51.890Z · comments (45)
Some of my disagreements with List of Lethalities
TurnTrout · 2023-01-24T00:25:28.075Z · comments (7)
Infohazards vs Fork Hazards
jimrandomh · 2023-01-05T09:45:28.065Z · comments (16)
AGI safety field building projects I’d like to see
Severin T. Seehrich (sts) · 2023-01-19T22:40:37.284Z · comments (27)
[link] Investing for a World Transformed by AI
PeterMcCluskey · 2023-01-01T02:47:06.004Z · comments (19)
How we could stumble into AI catastrophe
HoldenKarnofsky · 2023-01-13T16:20:05.745Z · comments (18)
Simulacra are Things
janus · 2023-01-08T23:03:26.052Z · comments (7)
[link] Tracr: Compiled Transformers as a Laboratory for Interpretability | DeepMind
DragonGod · 2023-01-13T16:53:10.279Z · comments (12)
Announcing aisafety.training
JJ Hepburn (jj-hepburn) · 2023-01-21T01:01:40.580Z · comments (4)
[link] Spooky action at a distance in the loss landscape
Jesse Hoogland (jhoogland) · 2023-01-28T00:22:46.506Z · comments (4)
LW Filter Tags (Rationality/World Modeling now promoted in Latest Posts)
Ruby · 2023-01-28T22:14:32.371Z · comments (4)
Escape Velocity from Bullshit Jobs
Zvi · 2023-01-10T14:30:00.828Z · comments (18)
Movie Review: Megan
Zvi · 2023-01-23T12:50:00.873Z · comments (19)
Assigning Praise and Blame: Decoupling Epistemology and Decision Theory
adamShimi · 2023-01-27T18:16:43.025Z · comments (5)
Inverse Scaling Prize: Second Round Winners
Ian McKenzie (naimenz) · 2023-01-24T20:12:48.474Z · comments (17)
[link] [Linkpost] TIME article: DeepMind’s CEO Helped Take AI Mainstream. Now He’s Urging Caution
Akash (akash-wasil) · 2023-01-21T16:51:09.586Z · comments (2)
My first year in AI alignment
Alex_Altair · 2023-01-02T01:28:03.470Z · comments (10)
[link] Conversational canyons
Henrik Karlsson (henrik-karlsson) · 2023-01-04T18:55:04.386Z · comments (4)
[link] Evidence under Adversarial Conditions
PeterMcCluskey · 2023-01-09T16:21:07.890Z · comments (1)
Consider paying for literature or book reviews using bounties and dominant assurance contracts
Arjun Panickssery (arjun-panickssery) · 2023-01-15T03:56:07.110Z · comments (7)
My Advice for Incoming SERI MATS Scholars
Johannes C. Mayer (johannes-c-mayer) · 2023-01-03T19:25:38.678Z · comments (1)
[link] Announcing Cavendish Labs
derikk · 2023-01-19T20:15:09.035Z · comments (5)
Linear Algebra Done Right, Axler
David Udell · 2023-01-02T22:54:58.724Z · comments (6)
Dangers of deference
TsviBT · 2023-01-08T14:36:33.454Z · comments (5)
Gradient Filtering
Jozdien · 2023-01-18T20:09:20.869Z · comments (16)
Consequentialists: One-Way Pattern Traps
David Udell · 2023-01-16T20:48:56.967Z · comments (3)
What’s going on with ‘crunch time’?
rosehadshar · 2023-01-20T09:42:53.215Z · comments (6)
[link] formal alignment: what it is, and some proposals
Tamsin Leake (carado-1) · 2023-01-29T11:32:33.239Z · comments (3)
[link] Why you should learn sign language
Noah Topper (noah-topper) · 2023-01-18T17:03:24.090Z · comments (23)
[link] Paper: Superposition, Memorization, and Double Descent (Anthropic)
LawrenceC (LawChan) · 2023-01-05T17:54:37.575Z · comments (11)
Reflections on Deception & Generality in Scalable Oversight (Another OpenAI Alignment Review)
Shoshannah Tekofsky (DarkSym) · 2023-01-28T05:26:49.866Z · comments (7)
Thoughts on hardware / compute requirements for AGI
Steven Byrnes (steve2152) · 2023-01-24T14:03:39.190Z · comments (30)
How Likely is Losing a Google Account?
jefftk (jkaufman) · 2023-01-30T00:20:01.584Z · comments (11)
Critique of some recent philosophy of LLMs’ minds
Roman Leventov · 2023-01-20T12:53:38.477Z · comments (8)
Contra Common Knowledge
abramdemski · 2023-01-04T22:50:38.493Z · comments (31)
[question] Would it be good or bad for the US military to get involved in AI risk?
Grant Demaree (grant-demaree) · 2023-01-01T19:02:30.892Z · answers+comments (12)
11 heuristics for choosing (alignment) research projects
Akash (akash-wasil) · 2023-01-27T00:36:08.742Z · comments (5)
[Simulators seminar sequence] #1 Background & shared assumptions
Jan (jan-2) · 2023-01-02T23:48:50.298Z · comments (4)
Trying to isolate objectives: approaches toward high-level interpretability
Jozdien · 2023-01-09T18:33:18.682Z · comments (14)
Language models can generate superior text compared to their input
ChristianKl · 2023-01-17T10:57:10.260Z · comments (28)
[RFC] Possible ways to expand on "Discovering Latent Knowledge in Language Models Without Supervision".
gekaklam · 2023-01-25T19:03:16.218Z · comments (6)
[link] NYT: Google will “recalibrate” the risk of releasing AI due to competition with OpenAI
Michael Huang · 2023-01-22T08:38:46.886Z · comments (2)
Citability of Lesswrong and the Alignment Forum
Leon Lang (leon-lang) · 2023-01-08T22:12:02.046Z · comments (2)
How-to Transformer Mechanistic Interpretability—in 50 lines of code or less!
StefanHex (Stefan42) · 2023-01-24T18:45:01.003Z · comments (5)
VIRTUA: a novel about AI alignment
Karl von Wendt · 2023-01-12T09:37:21.528Z · comments (12)
[Crosspost] ACX 2022 Prediction Contest Results
Scott Alexander (Yvain) · 2023-01-24T06:56:33.101Z · comments (6)
How to eat potato chips while typing
KatjaGrace · 2023-01-03T11:50:05.816Z · comments (12)
← previous page (newer posts) · next page (older posts) →