LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Renormalization Roadmap
Lauren Greenspan (LaurenGreenspan) · 2025-03-31T20:34:16.352Z · comments (7)
[link] Gary Marcus now saying AI can't do things it can already do
Benjamin_Todd · 2025-02-09T12:24:11.954Z · comments (12)
Fuzzing LLMs sometimes makes them reveal their secrets
Fabien Roger (Fabien) · 2025-02-26T16:48:48.878Z · comments (13)
FrontierMath Score of o3-mini Much Lower Than Claimed
YafahEdelman (yafah-edelman-1) · 2025-03-17T22:41:06.527Z · comments (7)
[link] The first RCT for GLP-1 drugs and alcoholism isn't what we hoped
dynomight · 2025-02-20T22:30:07.536Z · comments (4)
[link] How to Corner Liars: A Miasma-Clearing Protocol
ymeskhout · 2025-02-27T17:18:36.028Z · comments (23)
Falsehoods you might believe about people who are at a rationalist meetup
Screwtape · 2025-02-01T23:32:50.398Z · comments (12)
[link] Sentinel's Global Risks Weekly Roundup #11/2025. Trump invokes Alien Enemies Act, Chinese invasion barges deployed in exercise.
NunoSempere (Radamantis) · 2025-03-17T19:34:01.850Z · comments (3)
[link] How Gay is the Vatican?
rba · 2025-04-06T21:27:50.530Z · comments (32)
"Think it Faster" worksheet
Raemon · 2025-02-08T22:02:27.697Z · comments (8)
Not all capabilities will be created equal: focus on strategically superhuman agents
benwr · 2025-02-13T01:24:46.084Z · comments (8)
[link] Softmax, Emmett Shear's new AI startup focused on "Organic Alignment"
Chipmonk · 2025-03-28T21:23:46.220Z · comments (1)
On OpenAI’s Safety and Alignment Philosophy
Zvi · 2025-03-05T14:00:07.302Z · comments (5)
Alignment faking CTFs: Apply to my MATS stream
joshc (joshua-clymer) · 2025-04-04T16:29:02.070Z · comments (0)
Escape from Alderaan I
lsusr · 2025-02-02T10:48:06.533Z · comments (2)
Solving willpower seems easier than solving aging
Yair Halberstadt (yair-halberstadt) · 2025-03-23T15:25:40.861Z · comments (28)
A sketch of an AI control safety case
Tomek Korbak (tomek-korbak) · 2025-01-30T17:28:47.992Z · comments (0)
On Google’s Safety Plan
Zvi · 2025-04-11T12:51:12.112Z · comments (6)
Socially Graceful Degradation
Screwtape · 2025-03-20T04:03:41.213Z · comments (9)
Do models know when they are being evaluated?
Govind Pimpale (govind-pimpale) · 2025-02-17T23:13:22.017Z · comments (3)
Map of AI Safety v2
Bryce Robertson (bryceerobertson) · 2025-04-15T13:04:40.993Z · comments (4)
Go Grok Yourself
Zvi · 2025-02-19T20:20:09.371Z · comments (2)
The EU Is Asking for Feedback on Frontier AI Regulation (Open to Global Experts)—This Post Breaks Down What’s at Stake for AI Safety
Katalina Hernandez (katalina-hernandez) · 2025-04-22T20:39:40.781Z · comments (12)
[link] OpenAI Alums, Nobel Laureates Urge Regulators to Save Company's Nonprofit Structure
garrison · 2025-04-23T23:01:27.641Z · comments (0)
Housing Roundup #11
Zvi · 2025-04-01T16:30:03.694Z · comments (1)
How I switched careers from software engineer to AI policy operations
Lucie Philippon (lucie-philippon) · 2025-04-13T06:37:33.507Z · comments (1)
On polytopes
Dmitry Vaintrob (dmitry-vaintrob) · 2025-01-25T13:56:35.681Z · comments (5)
What's Behind the SynBio Bust?
sarahconstantin · 2025-01-30T22:30:06.916Z · comments (8)
Consider showering
bohaska (Bohaska) · 2025-04-01T23:54:26.714Z · comments (16)
[Closed] Gauging Interest for a Learning-Theoretic Agenda Mentorship Programme
Vanessa Kosoy (vanessa-kosoy) · 2025-02-16T16:24:57.654Z · comments (5)
To be legible, evidence of misalignment probably has to be behavioral
ryan_greenblatt · 2025-04-15T18:14:53.022Z · comments (14)
The Manus Marketing Madness
Zvi · 2025-03-10T20:10:07.845Z · comments (0)
On Deliberative Alignment
Zvi · 2025-02-11T13:00:07.683Z · comments (1)
[link] Dario Amodei: On DeepSeek and Export Controls
Zach Stein-Perlman · 2025-01-29T17:15:18.986Z · comments (3)
On MAIM and Superintelligence Strategy
Zvi · 2025-03-14T12:30:07.451Z · comments (2)
Interpreting Complexity
Maxwell Adam (intern) · 2025-03-14T04:52:32.103Z · comments (8)
Notes on countermeasures for exploration hacking (aka sandbagging)
ryan_greenblatt · 2025-03-24T18:39:36.665Z · comments (6)
Childhood and Education #9: School is Hell
Zvi · 2025-03-07T12:40:05.324Z · comments (36)
OpenAI Responses API changes models' behavior
Jan Betley (jan-betley) · 2025-04-11T13:27:29.942Z · comments (6)
Gemini 2.5 is the New SoTA
Zvi · 2025-03-28T14:20:03.176Z · comments (1)
≤10-year Timelines Remain Unlikely Despite DeepSeek and o3
Rafael Harth (sil-ver) · 2025-02-13T19:21:35.392Z · comments (59)
Reframing AI Safety as a Neverending Institutional Challenge
scasper · 2025-03-23T00:13:48.614Z · comments (12)
My "infohazards small working group" Signal Chat may have encountered minor leaks
Linch · 2025-04-02T01:03:05.311Z · comments (0)
[link] Modifying LLM Beliefs with Synthetic Document Finetuning
RowanWang (KevinRoWang) · 2025-04-24T21:15:17.366Z · comments (2)
HPMOR Anniversary Parties: Coordination, Resources, and Discussion
Screwtape · 2025-03-11T01:30:41.177Z · comments (6)
[link] You should read Hobbes, Locke, Hume, and Mill via EarlyModernTexts.com
Arjun Panickssery (arjun-panickssery) · 2025-01-30T12:35:03.564Z · comments (3)
[link] Conference Report: Threshold 2030 - Modeling AI Economic Futures
Deric Cheng (deric-cheng) · 2025-02-24T18:56:51.682Z · comments (0)
On OpenAI’s Model Spec 2.0
Zvi · 2025-02-21T14:10:06.827Z · comments (4)
AI #110: Of Course You Know…
Zvi · 2025-04-03T13:10:05.674Z · comments (9)
DeepSeek Panic at the App Store
Zvi · 2025-01-28T19:30:07.555Z · comments (14)
← previous page (newer posts) · next page (older posts) →