LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

"Liquidity" vs "solvency" in bank runs (and some notes on Silicon Valley Bank)
rossry · 2023-03-12T09:16:45.630Z · comments (27)
The Overton Window widens: Examples of AI risk in the media
Akash (akash-wasil) · 2023-03-23T17:10:14.616Z · comments (24)
[link] Perplexity wins my AI race
Elizabeth (pktechgirl) · 2024-08-24T19:20:10.859Z · comments (12)
The Darwin Game - Rounds 0 to 10
lsusr · 2020-10-24T02:17:43.343Z · comments (34)
Danger, AI Scientist, Danger
Zvi · 2024-08-15T22:40:06.715Z · comments (9)
Relationship Advice Repository
Ruby · 2022-06-20T14:39:36.548Z · comments (36)
Beliefs and Disagreements about Automating Alignment Research
Ian McKenzie (naimenz) · 2022-08-24T18:37:00.419Z · comments (4)
My current framework for thinking about AGI timelines
zhukeepa · 2020-03-30T01:23:57.195Z · comments (5)
Understanding Conjecture: Notes from Connor Leahy interview
Akash (akash-wasil) · 2022-09-15T18:37:51.653Z · comments (23)
[link] My techno-optimism [By Vitalik Buterin]
habryka (habryka4) · 2023-11-27T23:53:35.859Z · comments (17)
Merry Christmas
lsusr · 2021-12-26T07:03:30.261Z · comments (16)
One Day Sooner
Screwtape · 2023-11-02T19:00:58.427Z · comments (8)
[link] What's Wrong with Social Science and How to Fix It: Reflections After Reading 2578 Papers
habryka (habryka4) · 2020-09-12T01:46:07.349Z · comments (22)
[link] Most smart and skilled people are outside of the EA/rationalist community: an analysis
titotal (lombertini) · 2024-07-12T12:13:56.215Z · comments (36)
Programmatic backdoors: DNNs can use SGD to run arbitrary stateful computation
Fabien Roger (Fabien) · 2023-10-23T16:37:45.611Z · comments (3)
2023 in AI predictions
jessicata (jessica.liu.taylor) · 2024-01-01T05:23:42.514Z · comments (35)
Imitative Generalisation (AKA 'Learning the Prior')
Beth Barnes (beth-barnes) · 2021-01-10T00:30:35.976Z · comments (15)
Oversight Misses 100% of Thoughts The AI Does Not Think
johnswentworth · 2022-08-12T16:30:24.060Z · comments (49)
[link] Announcing turntrout.com, my new digital home
TurnTrout · 2024-11-17T17:42:08.164Z · comments (24)
Me, Myself, and AI: the Situational Awareness Dataset (SAD) for LLMs
L Rudolf L (LRudL) · 2024-07-08T22:24:38.441Z · comments (36)
[question] How do you feel about LessWrong these days? [Open feedback thread]
jacobjacob · 2023-12-05T20:54:42.317Z · answers+comments (284)
Petrov Day Retrospective: 2022
Ruby · 2022-09-28T22:16:20.325Z · comments (41)
Consider your appetite for disagreements
Adam Zerner (adamzerner) · 2022-10-08T23:25:44.096Z · comments (18)
Demystifying "Alignment" through a Comic
milanrosko · 2024-06-09T08:24:22.454Z · comments (19)
Did Bengio and Tegmark lose a debate about AI x-risk against LeCun and Mitchell?
Karl von Wendt · 2023-06-25T16:59:49.173Z · comments (53)
Catching AIs red-handed
ryan_greenblatt · 2024-01-05T17:43:10.948Z · comments (27)
Gradient hacking
evhub · 2019-10-16T00:53:00.735Z · comments (39)
Avoid Unnecessarily Political Examples
Raemon · 2021-01-11T05:41:56.439Z · comments (42)
New LessWrong feature: Dialogue Matching
jacobjacob · 2023-11-16T21:27:16.763Z · comments (22)
[link] Aristocracy and Hostage Capital
Arjun Panickssery (arjun-panickssery) · 2025-01-08T19:38:47.104Z · comments (7)
Why I'm doing PauseAI
Joseph Miller (Josephm) · 2024-04-30T16:21:54.156Z · comments (16)
Funding is All You Need: Getting into Grad School by Hacking the NSF GRFP Fellowship
hapanin · 2022-09-22T21:39:15.399Z · comments (9)
Skills I'd like my collaborators to have
Raemon · 2024-02-09T08:20:37.686Z · comments (9)
"If You're Not a Holy Madman, You're Not Trying"
abramdemski · 2021-02-28T18:56:19.560Z · comments (26)
How could you possibly choose what an AI wants?
So8res · 2023-04-19T17:08:54.694Z · comments (19)
Effective Evil
lsusr · 2021-11-02T00:26:29.910Z · comments (7)
The first future and the best future
KatjaGrace · 2024-04-25T06:40:04.510Z · comments (12)
"No evidence" as a Valley of Bad Rationality
Adam Zerner (adamzerner) · 2020-03-28T23:45:44.927Z · comments (21)
Trying to disambiguate different questions about whether RLHF is “good”
Buck · 2022-12-14T04:03:27.081Z · comments (47)
200 Concrete Open Problems in Mechanistic Interpretability: Introduction
Neel Nanda (neel-nanda-1) · 2022-12-28T21:06:53.853Z · comments (0)
Scaling and evaluating sparse autoencoders
leogao · 2024-06-06T22:50:39.440Z · comments (6)
Should we publish mechanistic interpretability research?
Marius Hobbhahn (marius-hobbhahn) · 2023-04-21T16:19:40.514Z · comments (40)
[Crosspost] On Hreha On Behavioral Economics
Scott Alexander (Yvain) · 2021-08-31T18:14:39.075Z · comments (6)
I don't think MIRI "gave up"
Raemon · 2023-02-03T00:26:07.552Z · comments (64)
Shapley Value Attribution in Chain of Thought
leogao · 2023-04-14T05:56:18.208Z · comments (7)
A transcript of the TED talk by Eliezer Yudkowsky
Mikhail Samin (mikhail-samin) · 2023-07-12T12:12:34.399Z · comments (13)
[link] My emotional reaction to the current funding situation
Sam F. Brown (sam-4) · 2022-09-09T22:02:46.301Z · comments (36)
Picking Mentors For Research Programmes
Raymond D · 2023-11-10T13:01:14.197Z · comments (8)
On the future of language models
owencb · 2023-12-20T16:58:28.433Z · comments (17)
Caution when interpreting Deepmind's In-context RL paper
Sam Marks (samuel-marks) · 2022-11-01T02:42:06.766Z · comments (8)
← previous page (newer posts) · next page (older posts) →