LessWrong 2.0 Reader

View: New · Old · Top

next page (older posts) →

What are the plans for solving the inner alignment problem?
Leonard Holloway (literally-best) · 2025-01-17T21:45:28.330Z · comments (0)
Your AI Safety focus is downstream of your AGI timeline
Michael Flood (michael-flood) · 2025-01-17T21:24:11.913Z · comments (0)
Thoughts on the conservative assumptions in AI control
Buck · 2025-01-17T19:23:38.575Z · comments (0)
Timaeus is hiring researchers & engineers
Jesse Hoogland (jhoogland) · 2025-01-17T19:13:14.739Z · comments (1)
Model Amnesty Project
themis · 2025-01-17T18:53:26.092Z · comments (0)
Addressing doubts of AI progress
LDJ (luigi-d) · 2025-01-17T18:53:06.343Z · comments (0)
Playing Dixit with AI: How Well LLMs Detect 'Me-ness'
Mariia Koroliuk (mariia-koroliuk) · 2025-01-17T18:52:23.488Z · comments (0)
Doing a self-randomized study of the impacts of glycine on sleep (Science is hard)
thedissonance.net · 2025-01-17T18:49:30.989Z · comments (1)
[link] How sci-fi can have drama without dystopia or doomerism
jasoncrawford · 2025-01-17T15:22:00.414Z · comments (1)
[question] What do you mean with ‘alignment is solvable in principle’?
Remmelt (remmelt-ellen) · 2025-01-17T15:03:12.138Z · answers+comments (3)
Meta Pivots on Content Moderation
Zvi · 2025-01-17T14:20:06.727Z · comments (1)
Tax Price Gouging?
jefftk (jkaufman) · 2025-01-17T14:10:03.395Z · comments (9)
The quantum red pill or: They lied to you, we live in the (density) matrix
Dmitry Vaintrob (dmitry-vaintrob) · 2025-01-17T13:58:16.186Z · comments (7)
Bednets -- 4 longer malaria studies
Hzn · 2025-01-17T08:47:50.342Z · comments (0)
Patent Trolling to Save the World
Double · 2025-01-17T04:13:46.768Z · comments (4)
Call Booth External Monitor
jefftk (jkaufman) · 2025-01-17T03:10:02.820Z · comments (0)
[Cross-post] Welcome to the Essay Meta
davekasten · 2025-01-16T23:36:49.152Z · comments (2)
AI for Resolving Forecasting Questions: An Early Exploration
ozziegooen · 2025-01-16T21:41:45.968Z · comments (2)
Ensuring long-term preservation of human consciousness
Amy Minge · 2025-01-16T19:09:18.925Z · comments (0)
[question] How Do You Interpret the Goal of LessWrong and Its Community?
ashen8461 · 2025-01-16T19:08:43.749Z · answers+comments (2)
[link] Experts' AI timelines are longer than you have been told?
Vasco Grilo (vascoamaralgrilo) · 2025-01-16T18:03:18.958Z · comments (4)
Numberwang: LLMs Doing Autonomous Research, and a Call for Input
eggsyntax · 2025-01-16T17:20:37.552Z · comments (19)
Topological Debate Framework
lunatic_at_large · 2025-01-16T17:19:25.816Z · comments (3)
AI #99: Farewell to Biden
Zvi · 2025-01-16T14:20:05.768Z · comments (2)
Deceptive Alignment and Homuncularity
Oliver Sourbut · 2025-01-16T13:55:19.161Z · comments (12)
Introducing the WeirdML Benchmark
Håvard Tveit Ihle (havard-tveit-ihle) · 2025-01-16T11:38:17.056Z · comments (13)
[link] The Mathematical Reason You should have 9 Kids
Zero Contradictions · 2025-01-16T11:24:46.617Z · comments (7)
Replicators, Gods and Buddhist Cosmology
KristianRonn · 2025-01-16T10:51:14.268Z · comments (1)
Quantum without complication
Optimization Process · 2025-01-16T08:53:11.347Z · comments (2)
Permanents: much more than you wanted to know
Dmitry Vaintrob (dmitry-vaintrob) · 2025-01-16T08:04:00.510Z · comments (2)
[link] Gaming TruthfulQA: Simple Heuristics Exposed Dataset Weaknesses
TurnTrout · 2025-01-16T02:14:35.098Z · comments (3)
What Is The Alignment Problem?
johnswentworth · 2025-01-16T01:20:16.826Z · comments (27)
Improving Our Safety Cases Using Upper and Lower Bounds
Yonatan Cale (yonatan-cale-1) · 2025-01-16T00:01:49.043Z · comments (0)
Unregulated Peptides: Does BPC-157 hold its promises?
ChristianKl · 2025-01-15T23:36:16.351Z · comments (5)
New, improved multiple-choice TruthfulQA
Owain_Evans · 2025-01-15T23:32:09.202Z · comments (0)
The Difference Between Prediction Markets and Debate (Argument) Maps
Jamie Joyce (jamie-joyce) · 2025-01-15T23:19:53.216Z · comments (3)
A Novel Emergence of Meta-Awareness in LLM Fine-Tuning
rife (edgar-muniz) · 2025-01-15T22:59:46.321Z · comments (8)
Six Small Cohabitive Games
Screwtape · 2025-01-15T21:59:29.778Z · comments (4)
LLMs are really good at k-order thinking (where k is even)
charlieoneill (kingchucky211) · 2025-01-15T20:43:00.623Z · comments (0)
Everywhere I Look, I See Kat Woods
just_browsing · 2025-01-15T19:29:10.373Z · comments (16)
Playground and Willpower Problems
Emre (emre-2) · 2025-01-15T18:52:08.860Z · comments (0)
"Pick Two" AI Trilemma: Generality, Agency, Alignment.
Black Flag (robert-shala-1) · 2025-01-15T18:52:00.780Z · comments (0)
Myths about Nonduality and Science by Gary Weber
Vadim Golub (a schizophrenic mind) · 2025-01-15T18:33:33.783Z · comments (0)
Marx and the Machine
DAL · 2025-01-15T18:33:16.789Z · comments (2)
Code4Compassion 2025: a hackathon transforming animal advocacy through technology
superbeneficiary · 2025-01-15T18:31:35.185Z · comments (0)
Applications Open for the Cooperative AI Summer School 2025!
JesseClifton · 2025-01-15T18:16:26.700Z · comments (0)
List of AI safety papers from companies, 2023–2024
Zach Stein-Perlman · 2025-01-15T18:00:30.242Z · comments (0)
AI Alignment Meme Viruses
RationalDino · 2025-01-15T15:55:24.271Z · comments (0)
[link] Looking for humanness in the world wide social
Itay Dreyfus (itay-dreyfus) · 2025-01-15T14:50:54.966Z · comments (0)
On the OpenAI Economic Blueprint
Zvi · 2025-01-15T14:30:06.773Z · comments (0)
next page (older posts) →