LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

next page (older posts) →

Maximizing Communication, not Traffic
jefftk (jkaufman) · 2025-01-05T13:00:02.280Z · comments (7)
Capital Ownership Will Not Prevent Human Disempowerment
beren · 2025-01-05T06:00:23.095Z · comments (8)
How will we update about scheming?
ryan_greenblatt · 2025-01-06T20:21:52.281Z · comments (3)
Reasons for and against working on technical AI safety at a frontier AI lab
bilalchughtai (beelal) · 2025-01-05T14:49:53.529Z · comments (12)
What Indicators Should We Watch to Disambiguate AGI Timelines?
snewman · 2025-01-06T19:57:43.398Z · comments (17)
[link] "We know how to build AGI" - Sam Altman
Nikola Jurkovic (nikolaisalreadytaken) · 2025-01-06T02:05:05.134Z · comments (5)
[link] Testing for Scheming with Model Deletion
Guive (GAA) · 2025-01-07T01:54:13.550Z · comments (1)
Estimating the benefits of a new flu drug (BXM)
DirectedEvolution (AllAmericanBreakfast) · 2025-01-06T04:31:16.837Z · comments (2)
[link] Oppression and production are competing explanations for wealth inequality.
Benquo · 2025-01-05T14:13:15.398Z · comments (13)
Childhood and Education #8: Dealing with the Internet
Zvi · 2025-01-06T14:00:09.604Z · comments (4)
Alternative Cancer Care As Biohacking & Book Review: Surviving "Terminal" Cancer
DenizT · 2025-01-06T07:43:52.773Z · comments (4)
Building Big Science from the Bottom-Up: A Fractal Approach to AI Safety
Lauren Greenspan (LaurenGreenspan) · 2025-01-07T03:08:51.447Z · comments (1)
Definition of alignment science I like
quetzal_rainbow · 2025-01-06T20:40:38.187Z · comments (0)
Really radical empathy
MichaelStJules · 2025-01-06T17:46:31.269Z · comments (0)
[link] AI safety content you could create
Adam Jones (domdomegg) · 2025-01-06T15:35:56.167Z · comments (0)
Measuring Nonlinear Feature Interactions in Sparse Crosscoders [Project Proposal]
Jason Gross (jason-gross) · 2025-01-06T04:22:12.633Z · comments (0)
Incredibow
jefftk (jkaufman) · 2025-01-07T03:30:02.197Z · comments (1)
[question] Meal Replacements in 2025?
alkjash · 2025-01-06T15:37:25.041Z · answers+comments (8)
[link] Policymakers don't have access to paywalled articles
Adam Jones (domdomegg) · 2025-01-05T10:56:11.495Z · comments (4)
(My) self-referential reason to believe in free will
jacek (jacek-karwowski) · 2025-01-06T23:35:02.809Z · comments (4)
[link] You should delay engineering-heavy research in light of R&D automation
Daniel Paleka · 2025-01-07T02:11:11.501Z · comments (1)
Guilt, Shame, and Depravity
Benquo · 2025-01-07T01:16:00.273Z · comments (2)
A Ground-Level Perspective on Capacity Building in International Development
Sean Aubin (sean-aubin) · 2025-01-05T20:36:54.308Z · comments (1)
Orange and Strawberry Truffles
jefftk (jkaufman) · 2025-01-05T01:50:01.587Z · comments (1)
[question] Is "hidden complexity of wishes problem" solved?
Roman Malov · 2025-01-05T22:59:30.911Z · answers+comments (4)
AXRP Episode 38.4 - Shakeel Hashim on AI Journalism
DanielFilan · 2025-01-05T00:20:05.096Z · comments (0)
Latent Adversarial Training (LAT) Improves the Representation of Refusal
alexandraabbas · 2025-01-06T10:24:53.419Z · comments (2)
D&D.Sci Dungeonbuilding: the Dungeon Tournament Evaluation & Ruleset
aphyer · 2025-01-07T05:02:25.929Z · comments (0)
Turning up the Heat on Deceptively-Misaligned AI
J Bostock (Jemist) · 2025-01-07T00:13:28.191Z · comments (0)
[question] Is my distinctiveness evidence for being in a simulation?
AynonymousPrsn123 · 2025-01-06T21:27:13.280Z · answers+comments (25)
[link] How to Do a PhD (in AI Safety)
Lewis Hammond (lewis-hammond-1) · 2025-01-05T16:57:35.409Z · comments (0)
Generating Cognateful Sentences with Large Language Models
vkethana (vijay-k) · 2025-01-06T18:40:09.564Z · comments (0)
Speedrunning Rationality: Day II
aproteinengine · 2025-01-06T03:59:25.488Z · comments (3)
Meditation insights as phase shifts in your self-model
Jonas Hallgren · 2025-01-07T10:09:35.854Z · comments (0)
We Will Likely Go Extinct Before the Unemployment Rate Reaches 99%. How Could That Happen?
Koki (Koki Takeda) · 2025-01-06T21:29:48.647Z · comments (0)
[link] Independent research article analyzing consistent self-reports of experience in ChatGPT and Claude
rife (edgar-muniz) · 2025-01-06T17:34:01.505Z · comments (7)
[link] My Experience Biohacking
Vale · 2025-01-07T03:01:21.410Z · comments (0)
Why Linear AI Safety Hits a Wall and How Fractal Intelligence Unlocks Non-Linear Solutions
Andy E Williams (andy-e-williams) · 2025-01-05T17:08:06.734Z · comments (5)
Alleviating shrimp pain is immoral.
G Wood (geoffrey-wood) · 2025-01-07T07:28:49.432Z · comments (0)
[link] Chinese Researchers Crack ChatGPT: Replicating OpenAI’s Advanced AI Model
Evan_Gaensbauer · 2025-01-05T03:50:34.245Z · comments (1)
next page (older posts) →