LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

next page (older posts) →

[link] Douglas Hofstadter changes his mind on Deep Learning & AI risk (June 2023)?
gwern · 2023-07-03T00:48:47.131Z · comments (54)
Alignment Grantmaking is Funding-Limited Right Now
johnswentworth · 2023-07-19T16:49:08.811Z · comments (67)
Accidentally Load Bearing
jefftk (jkaufman) · 2023-07-13T16:10:00.806Z · comments (14)
Yes, It's Subjective, But Why All The Crabs?
johnswentworth · 2023-07-28T19:35:36.741Z · comments (15)
Self-driving car bets
paulfchristiano · 2023-07-29T18:10:01.112Z · comments (41)
Ways I Expect AI Regulation To Increase Extinction Risk
1a3orn · 2023-07-04T17:32:48.047Z · comments (32)
[link] Cultivating a state of mind where new ideas are born
Henrik Karlsson (henrik-karlsson) · 2023-07-27T09:16:42.566Z · comments (18)
Consciousness as a conflationary alliance term for intrinsically valued internal experiences
Andrew_Critch · 2023-07-10T08:09:48.881Z · comments (46)
Grant applications and grand narratives
Elizabeth (pktechgirl) · 2023-07-02T00:16:25.129Z · comments (20)
[link] [Linkpost] Introducing Superalignment
beren · 2023-07-05T18:23:18.419Z · comments (68)
Towards Developmental Interpretability
Jesse Hoogland (jhoogland) · 2023-07-12T19:33:44.788Z · comments (8)
Cryonics and Regret
MvB (martin-von-berg) · 2023-07-24T09:16:01.456Z · comments (34)
My "2.9 trauma limit"
Raemon · 2023-07-01T19:32:14.805Z · comments (31)
Jailbreaking GPT-4's code interpreter
nikola (nikolaisalreadytaken) · 2023-07-13T18:43:54.484Z · comments (22)
OpenAI Launches Superalignment Taskforce
Zvi · 2023-07-11T13:00:06.232Z · comments (40)
Rationality !== Winning
Raemon · 2023-07-24T02:53:59.764Z · comments (49)
Brain Efficiency Cannell Prize Contest Award Ceremony
Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-07-24T11:30:10.602Z · comments (12)
When can we trust model evaluations?
evhub · 2023-07-28T19:42:21.799Z · comments (9)
[link] The Goddess of Everything Else - The Animation
Writer · 2023-07-13T16:26:25.552Z · comments (4)
The Seeker’s Game – Vignettes from the Bay
Yulia · 2023-07-09T19:32:58.717Z · comments (18)
[link] Neuronpedia
Johnny Lin (hijohnnylin) · 2023-07-26T16:29:28.884Z · comments (51)
[link] Introducing Fatebook: the fastest way to make and track predictions
Adam B (adam-b) · 2023-07-11T15:28:13.798Z · comments (34)
How LLMs are and are not myopic
janus · 2023-07-25T02:19:44.949Z · comments (14)
[link] Even Superhuman Go AIs Have Surprising Failure Modes
AdamGleave · 2023-07-20T17:31:35.814Z · comments (21)
Why was the AI Alignment community so unprepared for this moment?
Ras1513 · 2023-07-15T00:26:29.769Z · comments (65)
Going Crazy and Getting Better Again
Evenstar · 2023-07-02T18:55:25.790Z · comments (10)
Reducing sycophancy and improving honesty via activation steering
Nina Rimsky (NinaR) · 2023-07-28T02:46:23.122Z · comments (16)
“Reframing Superintelligence” + LLMs + 4 years
Eric Drexler · 2023-07-10T13:42:09.739Z · comments (8)
QAPR 5: grokking is maybe not *that* big a deal?
Quintin Pope (quintin-pope) · 2023-07-23T20:14:33.405Z · comments (15)
[link] Winners of AI Alignment Awards Research Contest
Akash (akash-wasil) · 2023-07-13T16:14:38.243Z · comments (3)
[link] Introducing bayescalc.io
Adele Lopez (adele-lopez-1) · 2023-07-07T16:11:12.854Z · comments (29)
Ten Levels of AI Alignment Difficulty
Sammy Martin (SDM) · 2023-07-03T20:20:21.403Z · comments (12)
Measuring and Improving the Faithfulness of Model-Generated Reasoning
Ansh Radhakrishnan (anshuman-radhakrishnan-1) · 2023-07-18T16:36:34.473Z · comments (13)
[link] Priorities for the UK Foundation Models Taskforce
Andrea_Miotti (AndreaM) · 2023-07-21T15:23:34.029Z · comments (4)
Consider Joining the UK Foundation Model Taskforce
Zvi · 2023-07-10T13:50:05.097Z · comments (12)
Anthropic Observations
Zvi · 2023-07-25T12:50:03.178Z · comments (1)
A transcript of the TED talk by Eliezer Yudkowsky
Mikhail Samin (mikhail-samin) · 2023-07-12T12:12:34.399Z · comments (13)
Views on when AGI comes and on strategy to reduce existential risk
TsviBT · 2023-07-08T09:00:19.735Z · comments (31)
When Someone Tells You They're Lying, Believe Them
ymeskhout · 2023-07-14T00:31:48.168Z · comments (3)
Fixed Point: a love story
Richard_Ngo (ricraz) · 2023-07-08T13:56:54.807Z · comments (2)
Why it's so hard to talk about Consciousness
Rafael Harth (sil-ver) · 2023-07-02T15:56:05.188Z · comments (152)
Apollo Neuro Results
Elizabeth (pktechgirl) · 2023-07-30T18:40:05.213Z · comments (16)
[question] What Does LessWrong/EA Think of Human Intelligence Augmentation as of mid-2023?
lukemarks (marc/er) · 2023-07-08T11:42:38.625Z · answers+comments (28)
BCIs and the ecosystem of modular minds
beren · 2023-07-21T15:58:27.081Z · comments (14)
Meta-level adversarial evaluation of oversight techniques might allow robust measurement of their adequacy
Buck · 2023-07-26T17:02:56.456Z · comments (18)
Compute Thresholds: proposed rules to mitigate risk of a “lab leak” accident during AI training runs
davidad · 2023-07-22T18:09:03.816Z · comments (2)
Underwater Torture Chambers: The Horror Of Fish Farming
omnizoid · 2023-07-26T00:27:15.490Z · comments (49)
[UPDATE: deadline extended to July 24!] New wind in rationality’s sails: Applications for Epistea Residency 2023 are now open
Jana Meixnerová (Epistea) · 2023-07-11T11:02:28.705Z · comments (7)
Sapient Algorithms
Valentine · 2023-07-17T16:30:01.350Z · comments (15)
[link] A $10k retroactive grant for VaccinateCA
Austin Chen (austin-chen) · 2023-07-27T18:14:44.305Z · comments (0)
next page (older posts) →