LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

[link] Should AIs be Encouraged to Cooperate?
PeterMcCluskey · 2025-04-15T21:57:06.096Z · comments (2)
A Talmudic Rationalist Cautionary Tale
Noah Birnbaum (daniel-birnbaum) · 2025-04-15T04:11:16.972Z · comments (1)
[link] A response to OpenAI’s “How we think about safety and alignment”
Harlan · 2025-03-31T20:58:31.901Z · comments (0)
Moonlight Reflected
Jacob Falkovich (Jacobian) · 2025-04-07T15:35:11.708Z · comments (0)
[link] The Case For Geopolitical Financial Speculation
prue (prue0) · 2025-04-01T21:09:17.515Z · comments (0)
The world according to ChatGPT
Richard_Kennaway · 2025-04-07T13:44:43.781Z · comments (0)
[link] Seeking feedback on "MAD Chairs: A new tool to evaluate AI"
Chris Santos-Lang (chris-santos-lang) · 2025-04-02T03:04:43.182Z · comments (0)
Theories of Impact for Causality in AI Safety
alexisbellot (alexis-1) · 2025-04-11T20:16:37.571Z · comments (1)
What does Yann LeCun think about AGI? A summary of his talk, "Mathematical Obstacles on the Way to Human-Level AI"
Adam Jones (domdomegg) · 2025-04-05T12:21:25.024Z · comments (0)
[link] Calculus is about change
dkl9 · 2025-04-01T19:44:43.453Z · comments (1)
Host Keys and SSHing to EC2
jefftk (jkaufman) · 2025-04-17T15:10:29.139Z · comments (6)
What are good safety standards for open source AIs from China?
ChristianKl · 2025-04-12T13:06:16.663Z · comments (2)
Story Feedback Request: The Policy - Emergent Alignment, Recursive Cognition, and AGI Trajectories
queelius · 2025-03-31T11:08:21.667Z · comments (2)
Cheesecake Frosting
jefftk (jkaufman) · 2025-04-04T02:10:07.755Z · comments (9)
[question] How likely are the USA to decay and how will it influence the AI development?
StanislavKrym · 2025-04-12T04:42:27.604Z · answers+comments (0)
How to enjoy fail attempts without self-deception (technique)
YanLyutnev (YanLutnev) · 2025-03-30T13:49:23.793Z · comments (0)
Misinformation is the default, and information is the government telling you your tap water is safe to drink
danielechlin · 2025-04-07T22:28:18.158Z · comments (2)
[link] Grounded Ghosts in the Machine - Friston Blankets, Mirror Neurons, and the Quest for Cooperative AI
Davidmanheim · 2025-04-10T10:15:54.880Z · comments (0)
[link] The Care and Feeding of Mythological Intelligences
Jack (jack-3) · 2025-04-02T22:05:21.151Z · comments (0)
Risers for Foot Percussion
jefftk (jkaufman) · 2025-04-15T11:10:08.577Z · comments (2)
The Mirror Problem in AI: Why Language Models Say Whatever You Want
RobT · 2025-04-15T18:40:02.793Z · comments (2)
Coupling for Decouplers — Intro
Jacob Falkovich (Jacobian) · 2025-04-07T15:12:26.892Z · comments (0)
Suggesting some revisions to Graham's hierarchy of disagreement
Sniffnoy · 2025-04-02T22:25:17.267Z · comments (2)
Nuanced Models for the Influence of Information
ozziegooen · 2025-04-10T18:28:34.082Z · comments (0)
[link] Human-level is not the limit
Vishakha (vishakha-agrawal) · 2025-04-16T08:33:15.498Z · comments (2)
[link] Paper Highlights, March '25
gasteigerjo · 2025-04-07T20:17:42.944Z · comments (0)
Building Communities Beyond the Bay
Lucie Philippon (lucie-philippon) · 2025-04-01T22:07:16.288Z · comments (2)
[Rockville] Rationalist Shabbat
maia · 2025-04-18T15:38:30.650Z · comments (0)
MATS is hiring!
Ryan Kidd (ryankidd44) · 2025-04-08T20:45:15.280Z · comments (0)
Comments on Karma systems
Arturo Macias (arturo-macias) · 2025-04-01T12:53:16.303Z · comments (2)
Yeshua's Basilisk
Alex Beyman (alexbeyman) · 2025-03-29T18:11:50.535Z · comments (1)
What empirical research directions has Eliezer commented positively on?
Chris_Leong · 2025-04-15T08:53:41.677Z · comments (1)
Linkpost to a Summary of "Imagining and building wise machines: The centrality of AI metacognition" by Johnson, Karimi, Bengio, et al.
Chris_Leong · 2025-04-10T11:54:37.484Z · comments (0)
[link] Conditional Forecasting as Model Parameterization
Molly (hickman-santini) · 2025-04-18T02:35:42.110Z · comments (0)
[Research sprint] Single-model crosscoder feature ablation and steering
Thomas Read (thjread) · 2025-04-06T14:42:30.357Z · comments (0)
Mass Exposure Paradox
max-sixty · 2025-04-16T20:18:00.492Z · comments (0)
Breaking down the MEAT of Alignment
JasonBrown · 2025-04-07T08:47:22.080Z · comments (2)
An Optimistic 2027 Timeline
Yitz (yitz) · 2025-04-06T16:39:36.554Z · comments (13)
I Have No Mouth but I Must Speak
Jack (jack-3) · 2025-04-05T07:42:54.424Z · comments (8)
0 Motivation Mapping through Information Theory
P. João (gabriel-brito) · 2025-04-18T00:53:34.360Z · comments (0)
[link] EA Reflections on my Military Career
TomGardiner (HorusXVI) · 2025-04-10T19:01:42.844Z · comments (0)
[link] Epoch AI is hiring a CTO!
merilalama · 2025-04-02T20:29:29.362Z · comments (0)
Commitment Races are a technical problem ASI can easily solve
Knight Lee (Max Lee) · 2025-04-12T22:22:47.790Z · comments (6)
The case for creating unaligned superintelligence
Yair Halberstadt (yair-halberstadt) · 2025-04-02T06:47:41.934Z · comments (0)
For Policy’s Sake: Why We Must Distinguish AI Safety from AI Security in Regulatory Governance
Katalina Hernandez (katalina-hernandez) · 2025-04-04T09:16:20.712Z · comments (11)
Enumerating objects a model "knows" using entity-detection features.
Alex Gibson · 2025-03-30T16:58:01.957Z · comments (2)
Arguing all sides with ChatGPT 4.5
Richard_Kennaway · 2025-04-07T13:10:11.562Z · comments (0)
[link] AISN #51: AI Frontiers
Corin Katzke (corin-katzke) · 2025-04-15T16:01:56.701Z · comments (1)
How Logic "Really" Works: An Engineering Perspective
Daniil Strizhov (mila-dolontaeva) · 2025-04-16T05:34:09.443Z · comments (0)
An idea for avoiding neuralese architectures
Knight Lee (Max Lee) · 2025-04-03T22:23:21.653Z · comments (2)
← previous page (newer posts) · next page (older posts) →