LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Public-facing Censorship Is Safety Theater, Causing Reputational Damage
Yitz (yitz) · 2022-09-23T05:08:14.149Z · comments (42)
Advice for newly busy people
Severin T. Seehrich (sts) · 2023-05-11T16:46:15.313Z · comments (3)
[link] Alcohol, health, and the ruthless logic of the Asian flush
dynomight · 2021-06-04T18:14:08.797Z · comments (45)
[link] Moral Reality Check (a short story)
jessicata (jessica.liu.taylor) · 2023-11-26T05:03:18.254Z · comments (45)
OpenAI #10: Reflections
Zvi · 2025-01-07T17:00:07.348Z · comments (7)
A Rocket–Interpretability Analogy
plex (ete) · 2024-10-21T13:55:18.184Z · comments (31)
Reply to Eliezer on Biological Anchors
HoldenKarnofsky · 2021-12-23T16:15:43.508Z · comments (46)
[Interim research report] Taking features out of superposition with sparse autoencoders
Lee Sharkey (Lee_Sharkey) · 2022-12-13T15:41:48.685Z · comments (23)
the scaling “inconsistency”: openAI’s new insight
nostalgebraist · 2020-11-07T07:40:06.548Z · comments (14)
[link] Dan Luu on "You can only communicate one top priority"
Raemon · 2023-03-18T18:55:09.998Z · comments (18)
On Devin
Zvi · 2024-03-18T13:20:04.779Z · comments (34)
At 87, Pearl is still able to change his mind
rotatingpaguro · 2023-10-18T04:46:29.339Z · comments (15)
There are no coherence theorems
Dan H (dan-hendrycks) · 2023-02-20T21:25:48.478Z · comments (130)
A Year of AI Increasing AI Progress
TW123 (ThomasWoodside) · 2022-12-30T02:09:39.458Z · comments (3)
Finite Factored Sets
Scott Garrabrant · 2021-05-23T20:52:48.575Z · comments (95)
Remarks 1–18 on GPT (compressed)
Cleo Nardo (strawberry calm) · 2023-03-20T22:27:26.277Z · comments (35)
Comments on OpenAI's "Planning for AGI and beyond"
So8res · 2023-03-03T23:01:29.665Z · comments (2)
The metaphor you want is "color blindness," not "blind spot."
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2022-02-14T00:28:55.119Z · comments (17)
[link] Why has nuclear power been a flop?
jasoncrawford · 2021-04-16T16:49:15.789Z · comments (50)
Moral public goods
paulfchristiano · 2020-01-26T00:10:01.803Z · comments (74)
All Possible Views About Humanity's Future Are Wild
HoldenKarnofsky · 2021-09-03T20:19:06.453Z · comments (37)
Could a superintelligence deduce general relativity from a falling apple? An investigation
titotal (lombertini) · 2023-04-23T12:49:44.288Z · comments (39)
Selection Has A Quality Ceiling
johnswentworth · 2021-06-02T18:25:54.432Z · comments (18)
AI coordination needs clear wins
evhub · 2022-09-01T23:41:48.334Z · comments (16)
Leading The Parade
johnswentworth · 2024-01-31T22:39:56.499Z · comments (31)
Reshaping the AI Industry
Thane Ruthenis · 2022-05-29T22:54:31.582Z · comments (35)
Subskills of "Listening to Wisdom"
Raemon · 2024-12-09T03:01:18.706Z · comments (29)
Discussion: Challenges with Unsupervised LLM Knowledge Discovery
Seb Farquhar · 2023-12-18T11:58:39.379Z · comments (21)
K-complexity is silly; use cross-entropy instead
So8res · 2022-12-20T23:06:27.131Z · comments (54)
Maximizing Communication, not Traffic
jefftk (jkaufman) · 2025-01-05T13:00:02.280Z · comments (10)
OpenAI o1
Zach Stein-Perlman · 2024-09-12T17:30:31.958Z · comments (41)
Some (problematic) aesthetics of what constitutes good work in academia
Steven Byrnes (steve2152) · 2024-03-11T17:47:28.835Z · comments (12)
Use Normal Predictions
Jan Christian Refsgaard (jan-christian-refsgaard) · 2022-01-09T15:01:44.934Z · comments (67)
Inner and outer alignment decompose one hard problem into two extremely hard problems
TurnTrout · 2022-12-02T02:43:20.915Z · comments (22)
Give it a google
Adam Zerner (adamzerner) · 2020-12-29T05:30:39.133Z · comments (28)
[link] "Heretical Thoughts on AI" by Eli Dourado
DragonGod · 2023-01-19T16:11:56.567Z · comments (38)
The U.S. is becoming less stable
lc · 2023-08-18T21:13:11.909Z · comments (68)
An Intuitive Guide to Garrabrant Induction
Mark Xu (mark-xu) · 2021-06-03T22:21:41.877Z · comments (20)
Repeal the Jones Act of 1920
Zvi · 2024-11-27T15:00:06.801Z · comments (23)
Why I’m not into the Free Energy Principle
Steven Byrnes (steve2152) · 2023-03-02T19:27:52.309Z · comments (49)
Authorities and Amateurs
jefftk (jkaufman) · 2020-03-25T03:40:02.208Z · comments (30)
6 non-obvious mental health issues specific to AI safety
Igor Ivanov (igor-ivanov) · 2023-08-18T15:46:09.938Z · comments (24)
Does davidad's uploading moonshot work?
Bird Concept (jacobjacob) · 2023-11-03T02:21:51.720Z · comments (35)
[link] Decomposing Agency — capabilities without desires
owencb · 2024-07-11T09:38:48.509Z · comments (32)
The "public debate" about AI is confusing for the general public and for policymakers because it is a three-sided debate
Adam David Long (adam-david-long-1) · 2023-08-01T00:08:30.908Z · comments (30)
Activation space interpretability may be doomed
bilalchughtai (beelal) · 2025-01-08T12:49:38.421Z · comments (28)
Why it's so hard to talk about Consciousness
Rafael Harth (sil-ver) · 2023-07-02T15:56:05.188Z · comments (193)
Assessing Kurzweil predictions about 2019: the results
Stuart_Armstrong · 2020-05-06T13:36:18.788Z · comments (22)
POC || GTFO culture as partial antidote to alignment wordcelism
lc · 2023-03-15T10:21:47.037Z · comments (13)
Algorithmic Improvement Is Probably Faster Than Scaling Now
johnswentworth · 2023-06-06T02:57:33.700Z · comments (25)
← previous page (newer posts) · next page (older posts) →