LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Dating Roundup #1: This is Why You’re Single
Zvi · 2023-08-29T12:50:04.964Z · comments (27)
[question] What are the best arguments for/against AIs being "slightly 'nice'"?
Raemon · 2024-09-24T02:00:19.605Z · answers+comments (49)
GPT-o1
Zvi · 2024-09-16T13:40:06.236Z · comments (34)
Scalable oversight as a quantitative rather than qualitative problem
Buck · 2024-07-06T17:42:41.325Z · comments (11)
Reflections on Less Online
Error · 2024-07-07T03:49:44.534Z · comments (15)
[link] Anxiety vs. Depression
Sable · 2024-03-17T00:15:08.255Z · comments (35)
Apollo Neuro Results
Elizabeth (pktechgirl) · 2023-07-30T18:40:05.213Z · comments (17)
Highlights: Wentworth, Shah, and Murphy on "Retargeting the Search"
RobertM (T3t) · 2023-09-14T02:18:05.890Z · comments (4)
The "spelling miracle": GPT-3 spelling abilities and glitch tokens revisited
mwatkins · 2023-07-31T19:47:02.793Z · comments (29)
Addressing Feature Suppression in SAEs
Benjamin Wright (Benw8888) · 2024-02-16T18:32:51.927Z · comments (3)
[link] Linkpost: Rishi Sunak's Speech on AI (26th October)
bideup · 2023-10-27T11:57:46.575Z · comments (8)
[link] The Puritans would one-box: evidential decision theory in the 17th century
Jacob G-W (g-w1) · 2023-10-14T20:23:24.346Z · comments (5)
[link] Environmentalism in the United States Is Unusually Partisan
Jeffrey Heninger (jeffrey-heninger) · 2024-05-13T21:23:10.755Z · comments (26)
Rejecting Television
Declan Molony (declan-molony) · 2024-04-23T04:59:50.253Z · comments (10)
Natural Latents: The Concepts
johnswentworth · 2024-03-20T18:21:19.878Z · comments (18)
[Valence series] 2. Valence & Normativity
Steven Byrnes (steve2152) · 2023-12-07T16:43:49.919Z · comments (5)
[link] Dario Amodei’s prepared remarks from the UK AI Safety Summit, on Anthropic’s Responsible Scaling Policy
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-11-01T18:10:31.110Z · comments (1)
[link] "AI Safety for Fleshy Humans" an AI Safety explainer by Nicky Case
habryka (habryka4) · 2024-05-03T18:10:12.478Z · comments (10)
Fluent, Cruxy Predictions
Raemon · 2024-07-10T18:00:06.424Z · comments (11)
Newsom Vetoes SB 1047
Zvi · 2024-10-01T12:20:06.127Z · comments (6)
My checklist for publishing a blog post
Steven Byrnes (steve2152) · 2023-08-15T15:04:56.219Z · comments (6)
[link] [Paper] Stress-testing capability elicitation with password-locked models
Fabien Roger (Fabien) · 2024-06-04T14:52:50.204Z · comments (10)
The case for unlearning that removes information from LLM weights
Fabien Roger (Fabien) · 2024-10-14T14:08:04.775Z · comments (3)
[link] A Universal Emergent Decomposition of Retrieval Tasks in Language Models
Alexandre Variengien (alexandre-variengien) · 2023-12-19T11:52:27.354Z · comments (3)
A simple case for extreme inner misalignment
Richard_Ngo (ricraz) · 2024-07-13T15:40:37.518Z · comments (41)
MATS Winter 2023-24 Retrospective
utilistrutil · 2024-05-11T00:09:17.059Z · comments (28)
Some for-profit AI alignment org ideas
Eric Ho (eh42) · 2023-12-14T14:23:20.654Z · comments (19)
[link] Nietzsche's Morality in Plain English
Arjun Panickssery (arjun-panickssery) · 2023-12-04T00:57:42.839Z · comments (13)
[link] Hardshipification
Jonathan Moregård (JonathanMoregard) · 2024-05-28T20:02:29.709Z · comments (17)
[link] [Paper] AI Sandbagging: Language Models can Strategically Underperform on Evaluations
Teun van der Weij (teun-van-der-weij) · 2024-06-13T10:04:49.556Z · comments (10)
Update on the UK AI Taskforce & upcoming AI Safety Summit
Elliot Mckernon (elliot) · 2023-10-11T11:37:42.436Z · comments (2)
[link] Anthropic's Responsible Scaling Policy & Long-Term Benefit Trust
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-09-19T15:09:27.235Z · comments (23)
AI #51: Altman’s Ambition
Zvi · 2024-02-20T19:50:07.439Z · comments (5)
Retirement Accounts and Short Timelines
jefftk (jkaufman) · 2024-02-19T18:50:05.231Z · comments (35)
[Intuitive self-models] 1. Preliminaries
Steven Byrnes (steve2152) · 2024-09-19T13:45:27.976Z · comments (18)
[link] The Real Fanfic Is The Friends We Made Along The Way
Eneasz · 2023-10-18T19:21:40.431Z · comments (0)
A Crisper Explanation of Simulacrum Levels
Thane Ruthenis · 2023-12-23T22:13:52.286Z · comments (13)
[link] What are you getting paid in?
Austin Chen (austin-chen) · 2024-07-17T19:23:04.219Z · comments (14)
Sparse Autoencoders Work on Attention Layer Outputs
Connor Kissane (ckkissane) · 2024-01-16T00:26:14.767Z · comments (9)
New roles on my team: come build Open Phil's technical AI safety program with me!
Ajeya Cotra (ajeya-cotra) · 2023-10-19T16:47:59.701Z · comments (6)
Actually, Power Plants May Be an AI Training Bottleneck.
Lao Mein (derpherpize) · 2024-06-20T04:41:33.567Z · comments (13)
Untrusted smart models and trusted dumb models
Buck · 2023-11-04T03:06:38.001Z · comments (12)
Why you should be using a retinoid
GeneSmith · 2024-08-19T03:07:41.722Z · comments (57)
Live Theory Part 0: Taking Intelligence Seriously
Sahil · 2024-06-26T21:37:10.479Z · comments (3)
Decomposing independent generalizations in neural networks via Hessian analysis
Dmitry Vaintrob (dmitry-vaintrob) · 2023-08-14T17:04:40.071Z · comments (4)
Saying the quiet part out loud: trading off x-risk for personal immortality
disturbance · 2023-11-02T17:43:34.155Z · comments (89)
Release: Optimal Weave (P1): A Prototype Cohabitive Game
mako yass (MakoYass) · 2024-08-17T14:08:18.947Z · comments (21)
Stepping down as moderator on LW
Kaj_Sotala · 2023-08-14T10:46:58.163Z · comments (1)
The Good Life in the face of the apocalypse
Elizabeth (pktechgirl) · 2023-10-16T22:40:15.200Z · comments (8)
[link] Essay competition on the Automation of Wisdom and Philosophy — $25k in prizes
owencb · 2024-04-16T10:10:13.338Z · comments (12)
← previous page (newer posts) · next page (older posts) →