LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Chaos Induces Abstractions
johnswentworth · 2021-03-18T20:08:21.739Z · comments (13)
H5N1
Zvi · 2023-02-13T12:50:00.694Z · comments (1)
[link] Advice for Activists from the History of Environmentalism
Jeffrey Heninger (jeffrey-heninger) · 2024-05-16T18:40:02.064Z · comments (8)
On the Executive Order
Zvi · 2023-11-01T14:20:01.657Z · comments (4)
The Power to Demolish Bad Arguments
Liron · 2019-09-02T12:57:23.341Z · comments (83)
Interpreting Yudkowsky on Deep vs Shallow Knowledge
adamShimi · 2021-12-05T17:32:26.532Z · comments (32)
[link] Discovering Language Model Behaviors with Model-Written Evaluations
evhub · 2022-12-20T20:08:12.063Z · comments (34)
Covid 11/12: The Winds of Winter
Zvi · 2020-11-12T14:30:01.387Z · comments (64)
Giving calibrated time estimates can have social costs
Alex_Altair · 2022-04-03T21:23:46.590Z · comments (16)
Anti-Corruption Market
lsusr · 2022-04-01T12:57:31.697Z · comments (23)
The Filan Cabinet Podcast with Oliver Habryka - Transcript
MondSemmel · 2023-02-14T02:38:34.867Z · comments (9)
LLM Modularity: The Separability of Capabilities in Large Language Models
NickyP (Nicky) · 2023-03-26T21:57:03.445Z · comments (3)
Work dumber not smarter
lukehmiles (lcmgcd) · 2023-06-01T12:40:31.264Z · comments (17)
[link] Instant stone (just add water!)
jasoncrawford · 2019-11-13T22:33:39.903Z · comments (27)
Open Source Sparse Autoencoders for all Residual Stream Layers of GPT2-Small
Joseph Bloom (Jbloom) · 2024-02-02T06:54:53.392Z · comments (37)
Evaluations project @ ARC is hiring a researcher and a webdev/engineer
Beth Barnes (beth-barnes) · 2022-09-09T22:46:47.569Z · comments (7)
Kelly *is* (just) about logarithmic utility
abramdemski · 2021-03-01T20:02:08.300Z · comments (26)
My take on Vanessa Kosoy's take on AGI safety
Steven Byrnes (steve2152) · 2021-09-30T12:23:58.329Z · comments (10)
How To Raise Others’ Aspirations in 17 Easy Steps
chanamessinger (cmessinger) · 2022-01-06T00:14:15.651Z · comments (13)
Comment on "Propositions Concerning Digital Minds and Society"
Zack_M_Davis · 2022-07-10T05:48:51.013Z · comments (12)
A Contamination Theory of the Obesity Epidemic
Bob Baker · 2021-07-25T02:39:14.676Z · comments (50)
Relevance Norms; Or, Gricean Implicature Queers the Decoupling/Contextualizing Binary
Zack_M_Davis · 2019-11-22T06:18:59.497Z · comments (30)
Iron deficiencies are very bad and you should treat them
Elizabeth (pktechgirl) · 2023-01-12T09:10:01.240Z · comments (30)
Petrov Day 2021: Mutually Assured Destruction?
Ruby · 2021-09-22T01:04:26.314Z · comments (96)
[link] everything is okay
Tamsin Leake (carado-1) · 2022-08-23T09:20:33.250Z · comments (22)
2019 Review: Voting Results!
Raemon · 2021-02-01T03:10:19.284Z · comments (36)
Parameter Scaling Comes for RL, Maybe
1a3orn · 2023-01-24T13:55:46.324Z · comments (3)
DeepMind: The Podcast - Excerpts on AGI
WilliamKiely · 2022-04-07T22:09:22.300Z · comments (11)
What's up with "Responsible Scaling Policies"?
habryka (habryka4) · 2023-10-29T04:17:07.839Z · comments (8)
The Adventure: a new Utopia story
Stuart_Armstrong · 2020-02-05T16:50:42.909Z · comments (37)
Naive Hypotheses on AI Alignment
Shoshannah Tekofsky (DarkSym) · 2022-07-02T19:03:49.458Z · comments (29)
Learning-theoretic agenda reading list
Vanessa Kosoy (vanessa-kosoy) · 2023-11-09T17:25:35.046Z · comments (0)
Marriage, the Giving What We Can Pledge, and the damage caused by vague public commitments
Jeffrey Ladish (jeff-ladish) · 2022-07-11T19:38:42.468Z · comments (27)
[link] [Link] Why I’m optimistic about OpenAI’s alignment approach
janleike · 2022-12-05T22:51:15.769Z · comments (15)
[Book review] Gödel, Escher, Bach: an in-depth explainer
Sam Marks (samuel-marks) · 2021-09-29T19:03:20.234Z · comments (23)
[link] Advice for journalists
Nathan Young · 2024-10-07T16:46:40.929Z · comments (53)
Truth and Advantage: Response to a draft of "AI safety seems hard to measure"
So8res · 2023-03-22T03:36:02.945Z · comments (9)
Announcing the London Initiative for Safe AI (LISA)
James Fox · 2024-02-02T23:17:47.011Z · comments (0)
The Teacup Test
lsusr · 2022-10-08T04:25:16.461Z · comments (32)
Units of Exchange
CFAR!Duncan (CFAR 2017) · 2022-06-28T16:53:53.069Z · comments (28)
Geoff Hinton Quits Google
Adam Shai (adam-shai) · 2023-05-01T21:03:47.806Z · comments (14)
[link] CIV: a story
Richard_Ngo (ricraz) · 2024-06-15T22:36:50.415Z · comments (6)
Hedonic asymmetries
paulfchristiano · 2020-01-26T02:10:01.323Z · comments (22)
Trying to Keep the Garden Well
Tobias H (clearthis) · 2022-01-16T05:42:11.851Z · comments (5)
The Social Alignment Problem
irving (judith) · 2023-04-28T14:16:17.825Z · comments (13)
Reframing Impact
TurnTrout · 2019-09-20T19:03:27.898Z · comments (15)
Total horse takeover
KatjaGrace · 2019-11-05T00:10:01.319Z · comments (14)
"Publish or Perish" (a quick note on why you should try to make your work legible to existing academic communities)
David Scott Krueger (formerly: capybaralet) (capybaralet) · 2023-03-18T19:01:54.199Z · comments (48)
Truthseeking when your disagreements lie in moral philosophy
Elizabeth (pktechgirl) · 2023-10-10T00:00:04.130Z · comments (4)
Salvage Epistemology
jimrandomh · 2022-04-30T02:10:41.996Z · comments (119)
← previous page (newer posts) · next page (older posts) →