LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

[link] The Grapes of Hardness
adamShimi · 2025-03-11T21:01:14.963Z · comments (0)
[link] Progress links and short notes, 2025-03-03
jasoncrawford · 2025-03-04T15:20:35.619Z · comments (0)
Decision-Relevance of worlds and ADT implementations
Maxime Riché (maxime-riche) · 2025-03-06T16:57:42.966Z · comments (0)
[link] Progress links and short notes, 2025-02-17
jasoncrawford · 2025-02-17T19:18:29.422Z · comments (0)
[link] A different take on the Musk v OpenAI preliminary injunction order
TFD · 2025-03-11T12:46:23.497Z · comments (0)
Talking to laymen about AI development
David Steel · 2025-02-17T18:42:23.289Z · comments (0)
[link] METR: AI models can be dangerous before public deployment
UnofficialLinkpostBot (LinkpostBot) · 2025-02-26T20:19:08.640Z · comments (0)
Conditional Importance in Toy Models of Superposition
james__p · 2025-02-02T20:35:38.655Z · comments (3)
What is the best / most proper definition of "Feeling the AGI" there is?
Annapurna (jorge-velez) · 2025-03-04T20:13:40.946Z · comments (5)
[link] Progress links and short notes, 2025-03-10
jasoncrawford · 2025-03-10T20:27:39.901Z · comments (0)
[link] Reply to Vitalik on d/acc
samuelshadrach (xpostah) · 2025-03-05T18:55:55.340Z · comments (0)
Technical comparison of Deepseek, Novasky, S1, Helix, P0
Juliezhanggg · 2025-02-25T04:20:40.413Z · comments (0)
The Structure of Professional Revolutions
SebastianG (JohnBuridan) · 2025-02-09T13:23:01.059Z · comments (0)
Amplifying the Computational No-Coincidence Conjecture
glauberdebona · 2025-03-07T21:29:54.933Z · comments (6)
[link] Forecasting newsletter #3/2025: Long march through the institutions
NunoSempere (Radamantis) · 2025-03-07T18:17:42.513Z · comments (0)
Post-hoc reasoning in chain of thought
Kyle Cox (klye) · 2025-02-05T18:58:29.802Z · comments (0)
You should use Consumer Reports
KvmanThinking (avery-liu) · 2025-02-27T01:52:17.235Z · comments (5)
Universal AI Maximizes Variational Empowerment: New Insights into AGI Safety
Yusuke Hayashi (hayashiyus) · 2025-02-27T00:46:46.989Z · comments (0)
[link] AI Safety at the Frontier: Paper Highlights, February '25
gasteigerjo · 2025-03-03T22:09:37.845Z · comments (0)
A Hogwarts Guide to Citizenship
WillPetillo · 2025-03-11T05:50:02.768Z · comments (1)
[link] Cooperation for AI safety must transcend geopolitical interference
Matrice Jacobine · 2025-02-16T18:18:01.539Z · comments (6)
What working on AI safety taught me about B2B SaaS sales
purple fire (jack-edwards) · 2025-02-04T20:50:19.990Z · comments (12)
Do we want alignment faking?
Florian_Dietz · 2025-02-28T21:50:48.891Z · comments (4)
Exploring how OthelloGPT computes its world model
JMaar (jim-maar) · 2025-02-02T21:29:09.433Z · comments (0)
[link] The Dilemma’s Dilemma
James Stephen Brown (james-brown) · 2025-02-19T23:50:47.485Z · comments (11)
[link] (Anti)Aging 101
George3d6 · 2025-03-12T03:59:21.859Z · comments (1)
Comparing the effectiveness of top-down and bottom-up activation steering for bypassing refusal on harmful prompts
Ana Kapros (ana-kapros) · 2025-02-12T19:12:07.592Z · comments (0)
[link] NY State Has a New Frontier Model Bill (+quick takes)
henryj · 2025-03-05T19:29:02.219Z · comments (0)
Sleeping Beauty: an Accuracy-based Approach
glauberdebona · 2025-02-10T15:40:29.619Z · comments (2)
Camps Should List Bands
jefftk (jkaufman) · 2025-03-06T03:00:02.348Z · comments (0)
Nationwide Action Workshop: Contact Congress about AI safety!
Felix De Simone (BobusChilc) · 2025-02-24T19:36:09.084Z · comments (0)
THE ARCHIVE
Jason Reid (jason-reid) · 2025-02-17T01:12:41.486Z · comments (0)
The old memories tree
Yair Halberstadt (yair-halberstadt) · 2025-03-05T19:03:59.498Z · comments (1)
Make Superintelligence Loving
Davey Morse (davey-morse) · 2025-02-21T06:07:17.235Z · comments (9)
[link] Neural Scaling Laws Rooted in the Data Distribution
aribrill (Particleman) · 2025-02-20T21:22:10.306Z · comments (0)
[question] Should I Divest from AI?
OKlogic · 2025-02-10T03:29:33.582Z · answers+comments (4)
Beyond ELO: Rethinking Chess Skill as a Multidimensional Random Variable
Oliver Oswald (oliver-oswald) · 2025-02-10T19:19:36.233Z · comments (7)
Arguing for the Truth? An Inference-Only Study into AI Debate
denisemester · 2025-02-11T03:04:58.852Z · comments (0)
Not-yet-falsifiable beliefs?
Benjamin Hendricks (benjamin-hendricks) · 2025-03-02T14:11:07.121Z · comments (4)
Build a Metaculus Forecasting Bot in 30 Minutes: A Practical Guide
ChristianWilliams · 2025-02-22T03:52:14.753Z · comments (0)
[link] AI Safety at the Frontier: Paper Highlights, January '25
gasteigerjo · 2025-02-11T16:14:16.972Z · comments (0)
[link] You don't actually need a physical multiverse to explain anthropic fine-tuning.
Fraser · 2025-03-12T07:33:43.278Z · comments (3)
Retroactive If-Then Commitments
MichaelDickens · 2025-02-01T22:22:43.031Z · comments (0)
Intelligence Is Jagged
Adam Train (aetrain) · 2025-02-19T07:08:46.444Z · comments (1)
Have you actually tried raising the birth rate?
Yair Halberstadt (yair-halberstadt) · 2025-03-10T18:06:40.987Z · comments (5)
Closed-ended questions aren't as hard as you think
electroswing · 2025-02-19T03:53:11.855Z · comments (0)
[question] Name for Standard AI Caveat?
yrimon (yehuda-rimon) · 2025-02-26T07:07:16.523Z · answers+comments (5)
Utilitarian AI Alignment: Building a Moral Assistant with the Constitutional AI Method
Clément L · 2025-02-04T04:15:36.917Z · comments (1)
[question] Does human (mis)alignment pose a significant and imminent existential threat?
jr · 2025-02-23T10:03:40.269Z · answers+comments (3)
There are a lot of upcoming retreats/conferences between March and July (2025)
gergogaspar (gergo-gaspar) · 2025-02-18T09:30:30.258Z · comments (0)
← previous page (newer posts) · next page (older posts) →