LessWrong 2.0 Reader

View: New · Old · Top

next page (older posts) →

Why we’re still doing normal school
juliawise · 2025-06-14T12:40:54.747Z · comments (0)
The Coaching Layer: Relational Intelligence for AI Safety
Priyanka Bharadwaj (priyanka-bharadwaj) · 2025-06-14T12:14:50.204Z · comments (0)
What Caused the Fertility Collapse?
Zero Contradictions · 2025-06-14T07:15:42.839Z · comments (0)
Relocation triggers
denkenberger · 2025-06-14T06:36:32.209Z · comments (0)
Memory Decoding Journal Club: Neocortical synaptic engrams for remote contextual memories
Devin Ward (Carboncopies Foundation) · 2025-06-14T02:26:41.113Z · comments (0)
[question] How concerned are you about a fast takeoff due to a leap in hardware usage?
MichaelDickens · 2025-06-14T01:15:03.629Z · answers+comments (4)
[question] How could I tell someone that consciousness is not the primary concern of AI Safety?
Lysandre Terrisse · 2025-06-13T22:44:34.319Z · answers+comments (2)
[link] Debate experiments at The Curve, LessOnline and Manifest
Nathan Young · 2025-06-13T22:35:14.805Z · comments (2)
[link] Futarchy's fundamental flaw
dynomight · 2025-06-13T22:08:11.031Z · comments (0)
[link] The Pros and Cons of Being Among Your Tribe
Sable · 2025-06-13T21:41:52.631Z · comments (0)
Constraining Minds, Not Goals: A Structural Approach to AI Alignment
Johannes C. Mayer (johannes-c-mayer) · 2025-06-13T21:06:40.984Z · comments (0)
[link] The optimal level of optimization is suboptimal
ellifournier (contrathinking) · 2025-06-13T18:06:02.611Z · comments (1)
On Pruning an Overgrown Garden
Vaatzes · 2025-06-13T17:54:09.725Z · comments (0)
Learned helplessness about "teaching to the test"
Viliam · 2025-06-13T17:53:10.649Z · comments (12)
[link] Information-Dense Conference Badges
ozziegooen · 2025-06-13T17:52:00.625Z · comments (3)
The Superwisdom Thesis: Why Superintelligence Does Not Pose An Existential Threat
Max Abecassis (max@customplay.com) · 2025-06-13T17:35:32.304Z · comments (9)
The Boat Theft Theory of Consciousness
Lorec · 2025-06-13T16:38:21.608Z · comments (16)
Monthly Roundup #31: June 2025
Zvi · 2025-06-13T16:20:01.739Z · comments (1)
Unsupervised Elicitation of Language Models
jiaxin wen · 2025-06-13T16:17:37.580Z · comments (0)
Unsupervised Elicitation of Language Models
Jiaxin Wen · 2025-06-13T16:15:09.045Z · comments (0)
Lucky Omega Problem
Tapatakt · 2025-06-13T14:54:56.718Z · comments (2)
[link] Distillation Robustifies Unlearning
Bruce W. Lee (bruce-lee) · 2025-06-13T13:45:26.261Z · comments (8)
Self-Adapting Language Models (from MIT, arXiv preprint)
Person (person-1) · 2025-06-13T13:08:54.715Z · comments (1)
[link] Do Not Tile the Lightcone with Your Confused Ontology
Jan_Kulveit · 2025-06-13T12:45:23.325Z · comments (4)
Corporations as Paperclip/Profit Maximizers
busssard (ole@falanx.de) · 2025-06-13T10:55:22.051Z · comments (1)
My reflections on doing a research fellowship Draft
Gauraventh (aryangauravyadav) · 2025-06-13T10:47:41.245Z · comments (0)
4. Why existing approaches to cause prioritization are not robust to unawareness
Anthony DiGiovanni (antimonyanthony) · 2025-06-13T08:55:23.521Z · comments (0)
[question] Under what conditions should humans stop pursuing technical AI safety careers?
S. Alex Bradt · 2025-06-13T05:56:07.911Z · answers+comments (0)
[link] [linkpost] AI Alignment is About Culture, Not Control by JCorvinus
Milan W (weibac) · 2025-06-13T00:07:21.834Z · comments (7)
[link] Forecast AI 2027
ChristianWilliams · 2025-06-12T21:12:04.836Z · comments (0)
[link] CRMArena-Pro: Holistic Assessment of LLM Agents Across Diverse Business Scenarios and Interactions
Annapurna (jorge-velez) · 2025-06-12T19:53:49.978Z · comments (0)
When does training a model change its goals?
Vivek Hebbar (Vivek) · 2025-06-12T18:43:05.187Z · comments (1)
Restraining Factors in AI Alignment Systems
theophilus tabuke (theophilus-tabuke) · 2025-06-12T18:17:32.252Z · comments (1)
[link] Analysis of Automated Prompt Engineering for Forecasting
ChristianWilliams · 2025-06-12T15:49:48.998Z · comments (0)
AI #120: While o3 Turned Pro
Zvi · 2025-06-12T15:30:03.664Z · comments (1)
Towards mutually assured cooperation
mikko (morrel) · 2025-06-12T15:15:02.188Z · comments (0)
What If We Could Monitor Human Intent?
Saif Khan (saif-khan) · 2025-06-12T08:51:00.785Z · comments (6)
The Way of a Skeptic
Martin Sustrik (sustrik) · 2025-06-12T05:40:10.684Z · comments (2)
[question] When should you read a biography?
CstineSublime · 2025-06-12T05:19:13.412Z · answers+comments (2)
[link] An Easily Overlooked Post on the Automation of Wisdom and Philosophy
Chris_Leong · 2025-06-12T02:54:50.303Z · comments (0)
Maybe Social Anxiety Is Just You Failing At Mind Control
25Hour (aaron-kaufman) · 2025-06-11T23:49:44.329Z · comments (15)
OpenAI now has an RL API which is broadly accessible
ryan_greenblatt · 2025-06-11T23:39:30.340Z · comments (0)
[link] So You Want to Work at a Frontier AI Lab
Joe Rogero · 2025-06-11T23:11:21.945Z · comments (6)
[link] Commentary On The Turing Apocrypha
jdp · 2025-06-11T22:52:27.845Z · comments (0)
[question] My friend wants a good book recommendation to understand AI, AI safety, and the field, and probably the drama. He’s smart but non-technical and not keeping up with trends. Any recs?
JohnGreer · 2025-06-11T22:32:37.808Z · answers+comments (0)
[link] The Dunning-Dunning-Kruger-Kruger Effect
ellifournier (contrathinking) · 2025-06-11T21:02:52.027Z · comments (2)
A Revision to Market Monetarism: Individual Hoarding as Rational, Competition for Dollars as Zero-Sum?
Lorec · 2025-06-11T20:13:33.050Z · comments (0)
A Systematic Study Extending "Emergent Misalignment": Causal Effects of Fine-Tuning Data on Model Vulnerability
Zhijing Jin · 2025-06-11T19:30:17.553Z · comments (0)
The Dream of a Gentle Singularity
Zvi · 2025-06-11T19:30:01.219Z · comments (6)
Beware General Claims about “Generalizable Reasoning Capabilities” (of Modern AI Systems)
LawrenceC (LawChan) · 2025-06-11T19:27:33.648Z · comments (9)
next page (older posts) →