LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Concrete Reasons for Hope about AI
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-01-14T01:22:18.723Z · comments (13)
My AGI safety research—2024 review, ’25 plans
Steven Byrnes (steve2152) · 2024-12-31T21:05:19.037Z · comments (4)
The Power to Demolish Bad Arguments
Liron · 2019-09-02T12:57:23.341Z · comments (84)
[link] Advice for Activists from the History of Environmentalism
Jeffrey Heninger (jeffrey-heninger) · 2024-05-16T18:40:02.064Z · comments (8)
Parameter Scaling Comes for RL, Maybe
1a3orn · 2023-01-24T13:55:46.324Z · comments (3)
Bad at Arithmetic, Promising at Math
cohenmacaulay · 2022-12-18T05:40:37.088Z · comments (19)
AI #8: People Can Do Reasonable Things
Zvi · 2023-04-20T15:50:00.826Z · comments (16)
[link] How counting neutrons explains nuclear waste
jasoncrawford · 2021-05-30T18:10:01.468Z · comments (11)
Don't accelerate problems you're trying to solve
Andrea_Miotti (AndreaM) · 2023-02-15T18:11:30.595Z · comments (27)
[link] LessOnline (May 31—June 2, Berkeley, CA)
Ben Pace (Benito) · 2024-03-26T02:34:00.000Z · comments (24)
The Mountain Troll
lsusr · 2022-06-11T09:14:01.479Z · comments (26)
A Case for the Least Forgiving Take On Alignment
Thane Ruthenis · 2023-05-02T21:34:49.832Z · comments (84)
Behavioral red-teaming is unlikely to produce clear, strong evidence that models aren't scheming
Buck · 2024-10-10T13:36:53.810Z · comments (4)
On the Executive Order
Zvi · 2023-11-01T14:20:01.657Z · comments (4)
[link] Advice for journalists
Nathan Young · 2024-10-07T16:46:40.929Z · comments (53)
AI Alignment Research Engineer Accelerator (ARENA): call for applicants
CallumMcDougall (TheMcDouglas) · 2023-04-17T20:30:03.965Z · comments (9)
[link] Discovering Language Model Behaviors with Model-Written Evaluations
evhub · 2022-12-20T20:08:12.063Z · comments (34)
[link] Knowledge Neurons in Pretrained Transformers
evhub · 2021-05-17T22:54:50.494Z · comments (7)
The Teacup Test
lsusr · 2022-10-08T04:25:16.461Z · comments (32)
Productive Mistakes, Not Perfect Answers
adamShimi · 2022-04-07T16:41:50.290Z · comments (11)
I'm still mystified by the Born rule
So8res · 2021-03-04T02:35:32.301Z · comments (44)
Work dumber not smarter
lemonhope (lcmgcd) · 2023-06-01T12:40:31.264Z · comments (17)
Covid 11/12: The Winds of Winter
Zvi · 2020-11-12T14:30:01.387Z · comments (64)
Reflections on the cryonics sequence
mingyuan · 2021-02-03T01:17:13.556Z · comments (11)
Giving calibrated time estimates can have social costs
Alex_Altair · 2022-04-03T21:23:46.590Z · comments (16)
[link] Draining the swamp
jasoncrawford · 2020-01-28T21:37:03.542Z · comments (1)
2019 Review: Voting Results!
Raemon · 2021-02-01T03:10:19.284Z · comments (36)
Truthseeking when your disagreements lie in moral philosophy
Elizabeth (pktechgirl) · 2023-10-10T00:00:04.130Z · comments (4)
[link] The Intelligence Curse
lukedrago · 2025-01-03T19:07:43.493Z · comments (26)
A Contamination Theory of the Obesity Epidemic
Bob Baker · 2021-07-25T02:39:14.676Z · comments (50)
What's up with "Responsible Scaling Policies"?
habryka (habryka4) · 2023-10-29T04:17:07.839Z · comments (8)
How To Raise Others’ Aspirations in 17 Easy Steps
chanamessinger (cmessinger) · 2022-01-06T00:14:15.651Z · comments (13)
Units of Exchange
CFAR!Duncan (CFAR 2017) · 2022-06-28T16:53:53.069Z · comments (28)
DeepMind: The Podcast - Excerpts on AGI
WilliamKiely · 2022-04-07T22:09:22.300Z · comments (11)
The Prototypical Negotiation Game
johnswentworth · 2021-02-20T21:33:34.195Z · comments (16)
[link] I found >800 orthogonal "write code" steering vectors
Jacob G-W (g-w1) · 2024-07-15T19:06:17.636Z · comments (19)
Comment on "Propositions Concerning Digital Minds and Society"
Zack_M_Davis · 2022-07-10T05:48:51.013Z · comments (12)
Relevance Norms; Or, Gricean Implicature Queers the Decoupling/Contextualizing Binary
Zack_M_Davis · 2019-11-22T06:18:59.497Z · comments (30)
LLM Modularity: The Separability of Capabilities in Large Language Models
NickyP (Nicky) · 2023-03-26T21:57:03.445Z · comments (3)
Petrov Day 2021: Mutually Assured Destruction?
Ruby · 2021-09-22T01:04:26.314Z · comments (96)
[link] Instant stone (just add water!)
jasoncrawford · 2019-11-13T22:33:39.903Z · comments (27)
Sam Altman's sister, Annie Altman, claims Sam has severely abused her
pythagoras5015 (pl5015) · 2023-10-07T21:06:49.396Z · comments (107)
[link] Nobody’s on the ball on AGI alignment
leopold · 2023-03-29T17:40:36.250Z · comments (38)
Learning-theoretic agenda reading list
Vanessa Kosoy (vanessa-kosoy) · 2023-11-09T17:25:35.046Z · comments (1)
You can, in fact, bamboozle an unaligned AI into sparing your life
David Matolcsi (matolcsid) · 2024-09-29T16:59:43.942Z · comments (171)
Selective, Corrective, Structural: Three Ways of Making Social Systems Work
Said Achmiz (SaidAchmiz) · 2023-03-05T08:45:45.615Z · comments (13)
Kelly *is* (just) about logarithmic utility
abramdemski · 2021-03-01T20:02:08.300Z · comments (26)
[link] The Minority Coalition
Richard_Ngo (ricraz) · 2024-06-24T20:01:27.436Z · comments (7)
"Publish or Perish" (a quick note on why you should try to make your work legible to existing academic communities)
David Scott Krueger (formerly: capybaralet) (capybaralet) · 2023-03-18T19:01:54.199Z · comments (49)
Evaluations project @ ARC is hiring a researcher and a webdev/engineer
Beth Barnes (beth-barnes) · 2022-09-09T22:46:47.569Z · comments (7)
← previous page (newer posts) · next page (older posts) →