LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Speed running everyone through the bad alignment bingo. $5k bounty for a LW conversational agent
ArthurB · 2023-03-09T09:26:25.383Z · comments (33)
[question] How to Convince my Son that Drugs are Bad
concerned_dad · 2022-12-17T18:47:24.398Z · answers+comments (84)
More Is Different for AI
jsteinhardt · 2022-01-04T19:30:20.352Z · comments (24)
The Learning-Theoretic Agenda: Status 2023
Vanessa Kosoy (vanessa-kosoy) · 2023-04-19T05:21:29.177Z · comments (17)
How might we align transformative AI if it’s developed very soon?
HoldenKarnofsky · 2022-08-29T15:42:08.985Z · comments (55)
A review of Steven Pinker's new book on rationality
Matthew Barnett (matthew-barnett) · 2021-09-29T01:29:58.151Z · comments (43)
Developmental Stages of GPTs
orthonormal · 2020-07-26T22:03:19.588Z · comments (72)
The Curse Of The Counterfactual
pjeby · 2019-11-01T18:34:41.186Z · comments (35)
Graphical tensor notation for interpretability
Jordan Taylor (Nadroj) · 2023-10-04T08:04:33.341Z · comments (11)
The Parable of the Boy Who Cried 5% Chance of Wolf
KatWoods (ea247) · 2022-08-15T14:33:21.649Z · comments (24)
Another RadVac Testing Update
johnswentworth · 2021-03-23T17:29:10.741Z · comments (19)
Against GDP as a metric for timelines and takeoff speeds
Daniel Kokotajlo (daniel-kokotajlo) · 2020-12-29T17:42:24.788Z · comments (19)
Coordination Schemes Are Capital Investments
Raemon · 2021-09-06T23:27:28.384Z · comments (31)
Understanding Infra-Bayesianism: A Beginner-Friendly Video Series
Jack Parker · 2022-09-22T13:25:04.254Z · comments (6)
A revolution in philosophy: the rise of conceptual engineering
Suspended Reason (suspended-reason) · 2020-06-02T18:30:30.495Z · comments (50)
Going Crazy and Getting Better Again
Evenstar · 2023-07-02T18:55:25.790Z · comments (13)
[link] Anomalous tokens reveal the original identities of Instruct models
janus · 2023-02-09T01:30:56.609Z · comments (16)
[link] Why didn't we get the four-hour workday?
jasoncrawford · 2023-01-06T21:29:38.995Z · comments (34)
A descriptive, not prescriptive, overview of current AI Alignment Research
Jan (jan-2) · 2022-06-06T21:59:22.344Z · comments (21)
AI Timelines via Cumulative Optimization Power: Less Long, More Short
jacob_cannell · 2022-10-06T00:21:02.447Z · comments (33)
Working in Virtual Reality: A Review
ozziegooen · 2020-11-20T23:14:28.707Z · comments (40)
Resolve Cycles
CFAR!Duncan (CFAR 2017) · 2022-07-16T23:17:13.037Z · comments (8)
The Apprentice Thread
Zvi · 2021-06-17T13:10:01.175Z · comments (59)
Category Theory Without The Baggage
johnswentworth · 2020-02-03T20:03:13.586Z · comments (51)
“Pivotal Act” Intentions: Negative Consequences and Fallacious Arguments
Andrew_Critch · 2022-04-19T20:25:35.018Z · comments (55)
Takeoff speeds have a huge effect on what it means to work on AI x-risk
Buck · 2022-04-13T17:38:11.990Z · comments (27)
What good is G-factor if you're dumped in the woods? A field report from a camp counselor.
Hastings (hastings-greer) · 2024-01-12T13:17:23.829Z · comments (22)
ELK prize results
paulfchristiano · 2022-03-09T00:01:02.085Z · comments (50)
Full Transcript: Eliezer Yudkowsky on the Bankless podcast
remember · 2023-02-23T12:34:19.523Z · comments (89)
I'm Sorry Fluttershy
sapphire (deluks917) · 2021-05-22T20:09:27.342Z · comments (4)
Comments on Carlsmith's “Is power-seeking AI an existential risk?”
So8res · 2021-11-13T04:29:30.673Z · comments (15)
The theory-practice gap
Buck · 2021-09-17T22:51:46.307Z · comments (15)
Honoring Petrov Day on LessWrong, in 2019
Ben Pace (Benito) · 2019-09-26T09:10:27.783Z · comments (168)
My experience using financial commitments to overcome akrasia
William Howard (william-howard) · 2024-04-15T22:57:32.574Z · comments (33)
How will we update about scheming?
ryan_greenblatt · 2025-01-06T20:21:52.281Z · comments (17)
Cosmopolitan values don't come free
So8res · 2023-05-31T15:58:16.974Z · comments (83)
[link] EfficientZero: human ALE sample-efficiency w/MuZero+self-supervised
gwern · 2021-11-02T02:32:41.856Z · comments (52)
Announcing the LessWrong Curated Podcast
Ben Pace (Benito) · 2022-06-22T22:16:58.170Z · comments (27)
Inflection.ai is a major AGI lab
Nikola Jurkovic (nikolaisalreadytaken) · 2023-08-09T01:05:54.604Z · comments (13)
Comparing Anthropic's Dictionary Learning to Ours
Robert_AIZI · 2023-10-07T23:30:32.402Z · comments (8)
LeCun’s “A Path Towards Autonomous Machine Intelligence” has an unsolved technical alignment problem
Steven Byrnes (steve2152) · 2023-05-08T19:35:19.180Z · comments (37)
The Seeker’s Game – Vignettes from the Bay
Yulia · 2023-07-09T19:32:58.717Z · comments (19)
[link] AI-Written Critiques Help Humans Notice Flaws
paulfchristiano · 2022-06-25T17:22:56.959Z · comments (5)
Inner Alignment in Salt-Starved Rats
Steven Byrnes (steve2152) · 2020-11-19T02:40:10.232Z · comments (41)
Defending the non-central fallacy
Matthew Barnett (matthew-barnett) · 2021-03-09T21:42:17.068Z · comments (38)
Yudkowsky vs Hanson on FOOM: Whose Predictions Were Better?
1a3orn · 2023-06-01T19:36:48.351Z · comments (76)
[$10k bounty] Read and compile Robin Hanson’s best posts
Richard_Ngo (ricraz) · 2021-10-20T22:03:47.376Z · comments (29)
Beyond Blame Minimization
physicaleconomics · 2022-03-27T00:03:31.650Z · comments (47)
2020 AI Alignment Literature Review and Charity Comparison
Larks · 2020-12-21T15:27:19.303Z · comments (14)
Read the Roon
Zvi · 2024-03-05T13:50:04.967Z · comments (6)
← previous page (newer posts) · next page (older posts) →