LessWrong 2.0 Reader

View: New · Old · Top

next page (older posts) →

Seeking submissions for short AI-safety course proposals
Sergio (sergio-abriola) · 2022-12-01T00:32:40.816Z · comments (0)
Reestablishing Reliable Sources: A System for Tagging URLs
Riley Mueller (rileymueller) · 2022-12-01T02:27:18.629Z · comments (1)
Notes on Caution
David Gross (David_Gross) · 2022-12-01T03:05:21.490Z · comments (0)
SBF's comments on ethics are no surprise to virtue ethicists
c.trout (ctrout) · 2022-12-01T04:18:25.877Z · comments (30)
[link] Did ChatGPT just gaslight me?
ThomasW (ThomasWoodside) · 2022-12-01T05:41:46.560Z · comments (45)
Safe Development of Hacker-AI Countermeasures – What if we are too late?
Erland Wittkotter (Erland) · 2022-12-01T07:59:11.862Z · comments (0)
Theories of impact for Science of Deep Learning
Marius Hobbhahn (marius-hobbhahn) · 2022-12-01T14:39:46.062Z · comments (0)
Research request (alignment strategy): Deep dive on "making AI solve alignment for us"
JanB (JanBrauner) · 2022-12-01T14:55:23.569Z · comments (3)
[link] [LINK] - ChatGPT discussion
JanB (JanBrauner) · 2022-12-01T15:04:45.257Z · comments (8)
[link] ChatGPT: First Impressions
specbug (rishit-vora) · 2022-12-01T16:36:19.592Z · comments (2)
Covid 12/1/22: China Protests
Zvi · 2022-12-01T17:10:00.839Z · comments (2)
The Machine Stops (Chapter 9)
Justin Bullock (justin-bullock) · 2022-12-01T19:20:26.031Z · comments (0)
Finding gliders in the game of life
paulfchristiano · 2022-12-01T20:40:04.230Z · comments (7)
The Plan - 2022 Update
johnswentworth · 2022-12-01T20:43:50.516Z · comments (37)
The LessWrong 2021 Review: Intellectual Circle Expansion
Ruby · 2022-12-01T21:17:50.321Z · comments (55)
Re-Examining LayerNorm
Eric Winsor (EricWinsor) · 2022-12-01T22:20:23.542Z · comments (12)
Take 1: We're not going to reverse-engineer the AI.
Charlie Steiner · 2022-12-01T22:41:32.677Z · comments (4)
Playing with Aerial Photos
jefftk (jkaufman) · 2022-12-01T22:50:04.609Z · comments (0)
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger (RobbBB) · 2022-12-01T23:11:44.279Z · comments (33)
[link] Understanding goals in complex systems
Johannes C. Mayer (johannes-c-mayer) · 2022-12-01T23:49:49.321Z · comments (0)
Lumenators for very lazy British people
shakeelh · 2022-12-02T00:18:36.876Z · comments (3)
Against meta-ethical hedonism
Joe Carlsmith (joekc) · 2022-12-02T00:23:26.039Z · comments (4)
Quick look: cognitive damage from well-administered anesthesia
Elizabeth (pktechgirl) · 2022-12-02T00:40:01.344Z · comments (0)
Update on Harvard AI Safety Team and MIT AI Alignment
Xander Davies (xanderdavies) · 2022-12-02T00:56:45.596Z · comments (4)
[link] Mastering Stratego (Deepmind)
[deleted] · 2022-12-02T02:21:56.672Z · comments (0)
New Feature: Collaborative editing now supports logged-out users
RobertM (T3t) · 2022-12-02T02:41:52.297Z · comments (0)
Inner and outer alignment decompose one hard problem into two extremely hard problems
TurnTrout · 2022-12-02T02:43:20.915Z · comments (22)
Deconfusing Direct vs Amortised Optimization
beren · 2022-12-02T11:30:46.754Z · comments (17)
Jailbreaking ChatGPT on Release Day
Zvi · 2022-12-02T13:10:00.860Z · comments (77)
[question] Is ChatGPT rigth when advising to brush the tongue when brushing teeth?
ChristianKl · 2022-12-02T14:53:02.123Z · answers+comments (14)
[link] NeurIPS Safety & ChatGPT. MLAISU W48
Esben Kran (esben-kran) · 2022-12-02T15:50:16.938Z · comments (0)
[ASoT] Finetuning, RL, and GPT's world prior
Jozdien · 2022-12-02T16:33:41.018Z · comments (8)
Takeoff speeds, the chimps analogy, and the Cultural Intelligence Hypothesis
NickGabs · 2022-12-02T19:14:59.825Z · comments (2)
Apply for the ML Upskilling Winter Camp in Cambridge, UK [2-10 Jan]
hannah wing-yee (hannah-erlebach) · 2022-12-02T20:45:10.768Z · comments (0)
Brun's theorem and sieve theory
Ege Erdil (ege-erdil) · 2022-12-02T20:57:39.956Z · comments (1)
Three Fables of Magical Girls and Longtermism
Ulisse Mini (ulisse-mini) · 2022-12-02T22:01:30.225Z · comments (11)
Research Principles for 6 Months of AI Alignment Studies
Shoshannah Tekofsky (DarkSym) · 2022-12-02T22:55:17.165Z · comments (3)
Subsets and quotients in interpretability
Erik Jenner (ejenner) · 2022-12-02T23:13:34.204Z · comments (1)
D&D.Sci December 2022: The Boojumologist
abstractapplic · 2022-12-02T23:39:49.398Z · comments (9)
Take 2: Building tools to help build FAI is a legitimate strategy, but it's dual-use.
Charlie Steiner · 2022-12-03T00:54:03.059Z · comments (1)
Causal Scrubbing: a method for rigorously testing interpretability hypotheses [Redwood Research]
LawrenceC (LawChan) · 2022-12-03T00:58:36.973Z · comments (35)
Causal scrubbing: Appendix
LawrenceC (LawChan) · 2022-12-03T00:58:45.850Z · comments (4)
Causal scrubbing: results on a paren balance checker
LawrenceC (LawChan) · 2022-12-03T00:59:08.078Z · comments (2)
Causal scrubbing: results on induction heads
LawrenceC (LawChan) · 2022-12-03T00:59:18.327Z · comments (1)
Great Cryonics Survey of 2022
Mati_Roy (MathieuRoy) · 2022-12-03T05:10:14.536Z · comments (0)
MrBeast's Squid Game Tricked Me
lsusr · 2022-12-03T05:50:02.339Z · comments (1)
[question] Is school good or bad?
tailcalled · 2022-12-03T13:14:22.737Z · answers+comments (76)
Our 2022 Giving
jefftk (jkaufman) · 2022-12-03T15:40:01.678Z · comments (0)
Utilitarianism is the only option
aelwood · 2022-12-03T17:14:19.532Z · comments (7)
[link] Chat GPT's views on Metaphysics and Ethics
Cole Killian (cole-killian) · 2022-12-03T18:12:19.290Z · comments (3)
next page (older posts) →