LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Explaining the Twitter Postrat Scene
Jacob Falkovich (Jacobian) · 2022-04-05T22:23:27.125Z · comments (27)
"Zero Sum" is a misnomer.
abramdemski · 2020-09-30T18:25:30.603Z · comments (34)
[link] Philosophy of Therapy
DaystarEld · 2020-10-10T20:12:38.204Z · comments (27)
Omicron Variant Post #2
Zvi · 2021-11-29T16:30:01.368Z · comments (34)
Natural Latents: The Math
johnswentworth · 2023-12-27T19:03:01.923Z · comments (37)
[link] DontDoxScottAlexander.com - A Petition
Ben Pace (Benito) · 2020-06-25T05:44:50.050Z · comments (32)
Moloch and the sandpile catastrophe
Eric Raymond (eric-raymond) · 2022-04-02T15:35:12.552Z · comments (25)
Propagating Facts into Aesthetics
Raemon · 2019-12-19T04:09:17.816Z · comments (37)
Harms and possibilities of schooling
TsviBT · 2022-02-22T07:48:09.542Z · comments (38)
[link] Matt Levine on "Fraud is no fun without friends."
Raemon · 2021-01-19T18:23:20.614Z · comments (24)
[link] Paper: LLMs trained on “A is B” fail to learn “B is A”
lberglund (brglnd) · 2023-09-23T19:55:53.427Z · comments (74)
Land Ho!
Zvi · 2022-01-20T13:30:01.262Z · comments (4)
Taking the parameters which seem to matter and rotating them until they don't
Garrett Baker (D0TheMath) · 2022-08-26T18:26:47.667Z · comments (48)
Ten Levels of AI Alignment Difficulty
Sammy Martin (SDM) · 2023-07-03T20:20:21.403Z · comments (14)
Stampy's AI Safety Info soft launch
steven0461 · 2023-10-05T22:13:04.632Z · comments (9)
[link] The Alignment Problem: Machine Learning and Human Values
Rohin Shah (rohinmshah) · 2020-10-06T17:41:21.138Z · comments (7)
Convincing All Capability Researchers
Logan Riggs (elriggs) · 2022-04-08T17:40:25.488Z · comments (70)
Compendium of problems with RLHF
Charbel-Raphaël (charbel-raphael-segerie) · 2023-01-29T11:40:53.147Z · comments (16)
My Understanding of Paul Christiano's Iterated Amplification AI Safety Research Agenda
Chi Nguyen · 2020-08-15T20:02:00.205Z · comments (20)
[link] Responsible Scaling Policies Are Risk Management Done Wrong
simeon_c (WayZ) · 2023-10-25T23:46:34.247Z · comments (34)
How to Bounded Distrust
Zvi · 2023-01-09T13:10:00.942Z · comments (16)
Quintin's alignment papers roundup - week 1
Quintin Pope (quintin-pope) · 2022-09-10T06:39:01.773Z · comments (6)
Evidence of Learned Look-Ahead in a Chess-Playing Neural Network
Erik Jenner (ejenner) · 2024-06-04T15:50:47.475Z · comments (14)
Problem relaxation as a tactic
TurnTrout · 2020-04-22T23:44:42.398Z · comments (8)
Future ML Systems Will Be Qualitatively Different
jsteinhardt · 2022-01-11T19:50:11.377Z · comments (10)
Christiano, Cotra, and Yudkowsky on AI progress
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2021-11-25T16:45:32.482Z · comments (95)
Geometric Exploration, Arithmetic Exploitation
Scott Garrabrant · 2022-11-24T15:36:30.334Z · comments (4)
Utilitarianism Meets Egalitarianism
Scott Garrabrant · 2022-11-21T19:00:12.168Z · comments (16)
Revealing Intentionality In Language Models Through AdaVAE Guided Sampling
jdp · 2023-10-20T07:32:28.749Z · comments (15)
Perpetual Dickensian Poverty?
jefftk (jkaufman) · 2021-12-21T13:30:03.543Z · comments (18)
A Significant Portion of COVID-19 Transmission Is Presymptomatic
jimrandomh · 2020-03-14T05:52:33.734Z · comments (22)
RTFB: On the New Proposed CAIP AI Bill
Zvi · 2024-04-10T18:30:08.410Z · comments (14)
Late 2021 MIRI Conversations: AMA / Discussion
Rob Bensinger (RobbBB) · 2022-02-28T20:03:05.318Z · comments (199)
Why was the AI Alignment community so unprepared for this moment?
Ras1513 · 2023-07-15T00:26:29.769Z · comments (65)
Unwitting cult leaders
Kaj_Sotala · 2021-02-11T11:10:04.504Z · comments (9)
FHI paper published in Science: interventions against COVID-19
SoerenMind · 2020-12-16T21:19:00.441Z · comments (0)
Delta Strain: Fact Dump and Some Policy Takeaways
Connor_Flexman · 2021-07-28T03:38:34.455Z · comments (60)
Solving the Mechanistic Interpretability challenges: EIS VII Challenge 1
StefanHex (Stefan42) · 2023-05-09T19:41:10.528Z · comments (1)
Circuits in Superposition: Compressing many small neural networks into one
Lucius Bushnaq (Lblack) · 2024-10-14T13:06:14.596Z · comments (7)
Cup-Stacking Skills (or, Reflexive Involuntary Mental Motions)
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2021-10-11T07:16:45.950Z · comments (36)
Mnestics
Jarred Filmer (4thWayWastrel) · 2022-10-23T00:30:11.159Z · comments (5)
[question] How do we prepare for final crunch time?
Eli Tyre (elityre) · 2021-03-30T05:47:54.654Z · answers+comments (30)
AI #14: A Very Good Sentence
Zvi · 2023-06-01T21:30:04.548Z · comments (30)
The case against AI alignment
andrew sauer (andrew-sauer) · 2022-12-24T06:57:53.405Z · comments (110)
Narrative Syncing
AnnaSalamon · 2022-05-01T01:48:45.889Z · comments (48)
I Would Have Solved Alignment, But I Was Worried That Would Advance Timelines
307th · 2023-10-20T16:37:46.541Z · comments (33)
The Standard Analogy
Zack_M_Davis · 2024-06-03T17:15:42.327Z · comments (28)
AI catastrophes and rogue deployments
Buck · 2024-06-03T17:04:51.206Z · comments (16)
Unifying Bargaining Notions (2/2)
Diffractor · 2022-07-27T03:40:30.524Z · comments (19)
Don't leave your fingerprints on the future
So8res · 2022-10-08T00:35:35.430Z · comments (46)
← previous page (newer posts) · next page (older posts) →