LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

Christiano (ARC) and GA (Conjecture) Discuss Alignment Cruxes
Andrea_Miotti (AndreaM) · 2023-02-24T23:03:04.917Z · comments (7)
Retrospective on the 2022 Conjecture AI Discussions
Andrea_Miotti (AndreaM) · 2023-02-24T22:41:13.131Z · comments (5)
How popular is ChatGPT? Part 1: more popular than Taylor Swift
Harlan · 2023-02-24T22:30:04.340Z · comments (0)
Are you stably aligned?
Seth Herd · 2023-02-24T22:08:23.098Z · comments (0)
Puzzle Cycles
Screwtape · 2023-02-24T21:35:09.052Z · comments (2)
[link] Sam Altman: "Planning for AGI and beyond"
LawrenceC (LawChan) · 2023-02-24T20:28:00.430Z · comments (54)
A Proposed Test to Determine the Extent to Which Large Language Models Understand the Real World
Bruce G · 2023-02-24T20:20:22.582Z · comments (7)
[link] Meta "open sources" LMs competitive with Chinchilla, PaLM, and code-davinci-002 (Paper)
LawrenceC (LawChan) · 2023-02-24T19:57:24.402Z · comments (19)
[link] Relationship Orientations
DaystarEld · 2023-02-24T19:43:41.463Z · comments (1)
The alien simulation meme doesn't make sense
[deleted] · 2023-02-24T19:27:11.916Z · comments (1)
[link] Exit Duty Generator by Matti Häyry
Oldphan · 2023-02-24T18:35:58.502Z · comments (0)
2023 Stanford Existential Risks Conference
elizabethcooper · 2023-02-24T18:35:39.663Z · comments (0)
How major governments can help with the most important century
HoldenKarnofsky · 2023-02-24T18:20:08.530Z · comments (0)
Consent Isn't Always Enough
jefftk (jkaufman) · 2023-02-24T15:40:05.048Z · comments (16)
[question] Training for corrigability: obvious problems?
Ben Amitay (unicode-70) · 2023-02-24T14:02:38.420Z · answers+comments (6)
Death and Desperation
Ustice · 2023-02-24T12:43:36.259Z · comments (3)
[question] Are there rationality techniques similar to staring at the wall for 4 hours?
trevor (TrevorWiesinger) · 2023-02-24T11:48:45.944Z · answers+comments (8)
The fast takeoff motte/bailey
lc · 2023-02-24T07:11:10.392Z · comments (7)
AGI systems & humans will both need to solve the alignment problem
Jeffrey Ladish (jeff-ladish) · 2023-02-24T03:29:21.043Z · comments (14)
A poor but certain attempt to philosophically undermine the orthogonality of intelligence and aims
Jay95 · 2023-02-24T03:03:57.927Z · comments (1)
I wanna Gandalf here
Igor Timofeev (igor-timofeev-1) · 2023-02-24T01:22:06.964Z · comments (4)
[link] [Link] A community alert about Ziz
DanielFilan · 2023-02-24T00:06:00.027Z · comments (126)
Teleosemantics!
abramdemski · 2023-02-23T23:26:15.894Z · comments (26)
AI that shouldn't work, yet kind of does
Donald Hobson (donald-hobson) · 2023-02-23T23:18:55.194Z · comments (8)
The AGI Optimist’s Dilemma
kaputmi · 2023-02-23T20:20:22.507Z · comments (1)
Searching for a model's concepts by their shape – a theoretical framework
Kaarel (kh) · 2023-02-23T20:14:46.341Z · comments (0)
[link] Why I'm Skeptical of De-Extinction
Niko_McCarty (niko-2) · 2023-02-23T19:42:52.618Z · comments (1)
[question] What causes randomness?
lotsofquestions · 2023-02-23T18:50:31.315Z · answers+comments (12)
Somerville Roads Getting More Dangerous?
jefftk (jkaufman) · 2023-02-23T18:20:03.354Z · comments (1)
EIS XII: Summary
scasper · 2023-02-23T17:45:55.973Z · comments (0)
How to survive in an AGI cataclysm
RomanS · 2023-02-23T14:34:53.998Z · comments (3)
Covid 2/23/23: Your Best Possible Situation
Zvi · 2023-02-23T13:10:01.887Z · comments (9)
Full Transcript: Eliezer Yudkowsky on the Bankless podcast
remember · 2023-02-23T12:34:19.523Z · comments (89)
Automated Sandwiching & Quantifying Human-LLM Cooperation: ScaleOversight hackathon results
Esben Kran (esben-kran) · 2023-02-23T10:48:08.766Z · comments (0)
[question] How to estimate a pre-aligned value for a common discussion ground?
EL_File4138 · 2023-02-23T10:38:18.489Z · answers+comments (12)
Interpersonal alignment intuitions
TekhneMakre · 2023-02-23T09:37:22.603Z · comments (18)
[link] Hello, Elua.
Tamsin Leake (carado-1) · 2023-02-23T05:19:07.246Z · comments (18)
Big Mac Subsidy?
jefftk (jkaufman) · 2023-02-23T04:00:03.996Z · comments (24)
[question] What moral systems (e.g utilitarianism) are common among LessWrong users?
hollowing · 2023-02-23T03:33:05.811Z · answers+comments (9)
AGI is likely to be cautious
PonPonPon · 2023-02-23T01:16:02.296Z · comments (14)
Short Notes on Research Process
Shoshannah Tekofsky (DarkSym) · 2023-02-22T23:41:45.279Z · comments (0)
[link] Video/animation: Neel Nanda explains what mechanistic interpretability is
DanielFilan · 2023-02-22T22:42:45.054Z · comments (7)
A Telepathic Exam about AI and Consequentialism
alkexr · 2023-02-22T21:00:21.994Z · comments (4)
[question] Injecting noise to GPT to get multiple answers
bipolo · 2023-02-22T20:02:13.644Z · answers+comments (1)
EIS XI: Moving Forward
scasper · 2023-02-22T19:05:52.723Z · comments (2)
Building and Entertaining Couples
Jacob Falkovich (Jacobian) · 2023-02-22T19:02:24.928Z · comments (11)
[link] Can submarines swim?
jasoncrawford · 2023-02-22T18:48:18.530Z · comments (14)
Is there a ML agent that abandons it's utility function out-of-distribution without losing capabilities?
Christopher King (christopher-king) · 2023-02-22T16:49:01.190Z · comments (7)
The male AI alignment solution
TekhneMakre · 2023-02-22T16:34:12.414Z · comments (24)
[link] Progress links and tweets, 2023-02-22
jasoncrawford · 2023-02-22T16:23:56.159Z · comments (0)
← previous page (newer posts) · next page (older posts) →