LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Two arguments against longtermist thought experiments
momom2 (amaury-lorin) · 2024-11-02T10:22:11.311Z · comments (5)
[link] Levers for Biological Progress - A Response to "Machines of Loving Grace"
Niko_McCarty (niko-2) · 2024-11-01T16:35:08.221Z · comments (0)
New Funding Category Open in Foresight's AI Safety Grants
Allison Duettmann (allison-duettmann) · 2024-11-06T22:59:41.065Z · comments (0)
LifeKeeper Diaries: Exploring Misaligned AI Through Interactive Fiction
Tristan Tran (tristan-tran) · 2024-11-09T20:58:09.182Z · comments (5)
Current Attitudes Toward AI Provide Little Data Relevant to Attitudes Toward AGI
Seth Herd · 2024-11-12T18:23:53.533Z · comments (2)
2024 NYC Secular Solstice & Megameetup
Joe Rogero · 2024-11-12T17:46:18.674Z · comments (0)
AXRP Episode 38.0 - Zhijing Jin on LLMs, Causality, and Multi-Agent Systems
DanielFilan · 2024-11-14T07:00:06.977Z · comments (0)
[link] Mechanistic Interpretability of Llama 3.2 with Sparse Autoencoders
PaulPauls · 2024-11-24T05:45:20.124Z · comments (0)
Dance Differentiation
jefftk (jkaufman) · 2024-11-15T02:30:07.694Z · comments (0)
Text Posts from the Kids Group: 2018
jefftk (jkaufman) · 2024-11-23T12:50:05.325Z · comments (0)
[link] AI & wisdom 2: growth and amortised optimisation
L Rudolf L (LRudL) · 2024-10-28T21:07:39.449Z · comments (0)
Secular Solstice Songbook Update
jefftk (jkaufman) · 2024-11-17T17:30:07.404Z · comments (1)
What can we learn from insecure domains?
Logan Zoellner (logan-zoellner) · 2024-11-01T23:53:30.066Z · comments (21)
[question] How can we prevent AGI value drift?
Dakara (chess-ice) · 2024-11-20T18:19:24.375Z · answers+comments (4)
[link] What if muscle tension is sometimes signal jamming?
Chipmonk · 2024-11-04T21:08:47.800Z · comments (1)
[link] AI & wisdom 3: AI effects on amortised optimisation
L Rudolf L (LRudL) · 2024-10-28T21:08:56.604Z · comments (0)
[question] Why is Gemini telling the user to die?
Burny · 2024-11-18T01:44:12.583Z · answers+comments (1)
[link] The lying p value
kqr · 2024-11-12T06:12:59.934Z · comments (6)
Goal: Understand Intelligence
Johannes C. Mayer (johannes-c-mayer) · 2024-11-03T21:20:02.900Z · comments (19)
Registrations Open for 2024 NYC Secular Solstice & Megameetup
Joe Rogero · 2024-11-12T17:50:10.827Z · comments (0)
Curriculum of Ascension
andrew sauer (andrew-sauer) · 2024-11-07T23:54:18.983Z · comments (0)
GPT-4o Can In Some Cases Solve Moderately Complicated Captchas
dirk (abandon) · 2024-11-09T04:04:37.782Z · comments (2)
AXRP Episode 38.1 - Alan Chan on Agent Infrastructure
DanielFilan · 2024-11-16T23:30:09.098Z · comments (0)
A Case for Conscious Significance rather than Free Will.
James Stephen Brown (james-brown) · 2024-10-25T23:20:30.834Z · comments (2)
ML4Good (AI Safety Bootcamp) - Experience report
JanEbbing · 2024-11-05T01:18:43.554Z · comments (0)
The current state of RSPs
Zach Stein-Perlman · 2024-11-04T16:00:42.630Z · comments (0)
A Poem Is All You Need: Jailbreaking ChatGPT, Meta & More
Sharat Jacob Jacob (sharat-jacob-jacob) · 2024-10-29T12:41:30.337Z · comments (0)
Paraddictions: unreasonably compelling behaviors and their uses
Michael Cohn (michael-cohn) · 2024-11-22T20:53:59.479Z · comments (0)
Fundamental Uncertainty: Chapter 9 - How do we live with uncertainty?
Gordon Seidoh Worley (gworley) · 2024-11-07T18:15:45.049Z · comments (2)
[link] Anthropic - The case for targeted regulation
anaguma · 2024-11-05T07:07:48.174Z · comments (0)
[link] [Linkpost] Building Altruistic and Moral AI Agent with Brain-inspired Affective Empathy Mechanisms
Gunnar_Zarncke · 2024-11-04T10:15:35.550Z · comments (0)
Spooky Recommendation System Scaling
phdead · 2024-10-31T22:00:51.728Z · comments (0)
Updating the NAO Simulator
jefftk (jkaufman) · 2024-10-30T13:50:06.908Z · comments (0)
Don't Dismiss on Epistemics
ggex · 2024-11-19T00:44:05.329Z · comments (3)
[link] OpenAI’s cybersecurity is probably regulated by NIS Regulations
Adam Jones (domdomegg) · 2024-10-25T11:06:38.392Z · comments (2)
Substituting Talkbox for Breath Controller
jefftk (jkaufman) · 2024-10-27T19:10:03.768Z · comments (0)
Why We Wouldn't Build Aligned AI Even If We Could
Snowyiu · 2024-11-16T20:19:59.324Z · comments (6)
[question] What are some positive developments in AI safety in 2024?
Satron · 2024-11-15T10:32:39.541Z · answers+comments (5)
Prediction markets and Taxes
Edmund Nelson (edmund-nelson) · 2024-11-01T17:39:35.191Z · comments (7)
Fundamental Uncertainty: Epilogue
Gordon Seidoh Worley (gworley) · 2024-11-16T00:57:48.823Z · comments (0)
Reward Bases: A simple mechanism for adaptive acquisition of multiple reward type
Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2024-11-23T12:45:01.067Z · comments (0)
Aligning AI Safety Projects with a Republican Administration
Deric Cheng (deric-cheng) · 2024-11-21T22:12:27.502Z · comments (0)
Making a Pedalboard
jefftk (jkaufman) · 2024-10-25T00:10:09.149Z · comments (0)
Sideloading: creating a model of a person via LLM with very large prompt
avturchin · 2024-11-22T16:41:28.293Z · comments (4)
[question] Using hex to get murder advice from GPT-4o
Laurence Freeman (laurence-freeman) · 2024-11-13T18:30:23.475Z · answers+comments (5)
Expected Utility, Geometric Utility, and Other Equivalent Representations
StrivingForLegibility · 2024-11-20T23:28:21.826Z · comments (0)
Festival Stats 2024
jefftk (jkaufman) · 2024-11-12T02:00:04.831Z · comments (0)
[link] Proposing the Conditional AI Safety Treaty (linkpost TIME)
otto.barten (otto-barten) · 2024-11-15T13:59:01.050Z · comments (8)
[link] Book Review: Replacing Guilt - On Having Something to Fight For
Cole Killian (cole-killian) · 2024-11-03T19:47:35.093Z · comments (0)
[question] What are some good ways to form opinions on controversial subjects in the current and upcoming era?
notfnofn · 2024-10-27T14:33:53.960Z · answers+comments (21)
← previous page (newer posts) · next page (older posts) →