LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

Training Trace Priors and Speed Priors
Adam Jermyn (adam-jermyn) · 2022-06-26T18:07:08.746Z · comments (0)
My current take on Internal Family Systems “parts”
Kaj_Sotala · 2022-06-26T17:40:05.750Z · comments (10)
A Quick Ontology of Agreement
ravedon · 2022-06-26T17:39:25.315Z · comments (2)
Seven ways to become unstoppably agentic
[deleted] · 2022-06-26T17:39:08.424Z · comments (16)
Formalizing Deception
JamesH (AtlasOfCharts) · 2022-06-26T17:39:01.390Z · comments (2)
Dust Theory vs Ruliad
[deleted] · 2022-06-26T16:08:34.813Z · comments (0)
My cognitive inertia cycle
MSRayne · 2022-06-26T15:49:17.661Z · comments (4)
How do poor countries get rich: some theories
NathanBarnard · 2022-06-26T10:41:36.027Z · comments (2)
Child Contracting
jefftk (jkaufman) · 2022-06-26T02:30:04.904Z · comments (2)
Conditioning Generative Models
Adam Jermyn (adam-jermyn) · 2022-06-25T22:15:59.138Z · comments (18)
[link] Unforgivable
Novalis · 2022-06-25T20:57:47.297Z · comments (12)
[link] SunPJ in Alenia
FlorianH (florian-habermacher) · 2022-06-25T19:39:50.393Z · comments (19)
[question] Should any human enslave an AGI system?
AlignmentMirror · 2022-06-25T19:35:54.079Z · answers+comments (44)
Fundamental Uncertainty: Chapter 3 - Why don't we agree on what's right?
Gordon Seidoh Worley (gworley) · 2022-06-25T17:50:37.565Z · comments (21)
[question] How "should" counterfactual prediction markets work?
eapi (edward-pierzchalski) · 2022-06-25T17:44:58.497Z · answers+comments (6)
Conversation with Eliezer: What do you want the system to do?
Akash (akash-wasil) · 2022-06-25T17:36:14.145Z · comments (38)
[link] AI-Written Critiques Help Humans Notice Flaws
paulfchristiano · 2022-06-25T17:22:56.959Z · comments (5)
Some reflections on the LW community after several months of active engagement
M. Y. Zuo · 2022-06-25T17:04:16.233Z · comments (40)
[link] On The Spectrum, On The Guest List: (vii) The Marquee
party girl (party-girl) · 2022-06-25T16:54:59.809Z · comments (0)
Identification of Natural Modularity
Stephen Fowler (LosPolloFowler) · 2022-06-25T15:05:17.793Z · comments (3)
[LQ] Some Thoughts on Messaging Around AI Risk
DragonGod · 2022-06-25T13:53:26.833Z · comments (3)
[link] Quick Summaries of Two Papers on Kant and Game Theory
Erich_Grunewald · 2022-06-25T10:25:08.398Z · comments (2)
[question] Do you consider your current, non-superhuman self aligned with “humanity” already?
Rana Dexsin · 2022-06-25T04:15:08.088Z · answers+comments (19)
LW/ACX/EA Seattle summer meetup
Nikita Sokolsky (nikita-sokolsky) · 2022-06-24T23:30:31.366Z · comments (2)
Dependencies for AGI pessimism
Yitz (yitz) · 2022-06-24T22:25:03.049Z · comments (4)
[link] [Link] Childcare : what the science says
Gunnar_Zarncke · 2022-06-24T21:45:23.406Z · comments (4)
What if the best path for a person who wants to work on AGI alignment is to join Facebook or Google?
dbasch · 2022-06-24T21:23:01.758Z · comments (3)
[Link] Adversarially trained neural representations may already be as robust as corresponding biological neural representations
Gunnar_Zarncke · 2022-06-24T20:51:27.924Z · comments (9)
Updated Deference is not a strong argument against the utility uncertainty approach to alignment
Ivan Vendrov (ivan-vendrov) · 2022-06-24T19:32:17.547Z · comments (8)
[link] Cracks in the Wall, Part I: The Conscious
silo · 2022-06-24T18:29:15.231Z · comments (28)
[question] Do alignment concerns extend to powerful non-AI agents?
Ozyrus · 2022-06-24T18:26:22.737Z · answers+comments (13)
[link] Raphaël Millière on Generalization and Scaling Maximalism
Michaël Trazzi (mtrazzi) · 2022-06-24T18:18:10.503Z · comments (2)
Worked Examples of Shapley Values
lalaithion · 2022-06-24T17:13:43.095Z · comments (9)
Feature request: voting buttons at the bottom?
Oliver Sourbut · 2022-06-24T14:41:55.268Z · comments (12)
Intelligence in Commitment Races
David Udell · 2022-06-24T14:30:21.525Z · comments (7)
[link] Linkpost: Robin Hanson - Why Not Wait On AI Risk?
Yair Halberstadt (yair-halberstadt) · 2022-06-24T14:23:50.580Z · comments (14)
[question] "Science Cathedrals"
Alex Vermillion (tomcatfish) · 2022-06-24T03:30:12.854Z · answers+comments (9)
LessWrong Has Agree/Disagree Voting On All New Comment Threads
Ben Pace (Benito) · 2022-06-24T00:43:17.136Z · comments (217)
Book review: The Passenger by Lisa Lutz
KatjaGrace · 2022-06-23T23:10:19.626Z · comments (1)
20 Critiques of AI Safety That I Found on Twitter
dkirmani · 2022-06-23T19:23:20.013Z · comments (16)
[link] The Limits of Automation
milkandcigarettes · 2022-06-23T18:03:12.881Z · comments (1)
[question] Is CIRL a promising agenda?
Chris_Leong · 2022-06-23T17:12:51.213Z · answers+comments (16)
[Link] OpenAI: Learning to Play Minecraft with Video PreTraining (VPT)
Aryeh Englander (alenglander) · 2022-06-23T16:29:19.323Z · comments (3)
Half-baked AI Safety ideas thread
Aryeh Englander (alenglander) · 2022-06-23T16:11:54.207Z · comments (61)
Nonprofit Boards are Weird
HoldenKarnofsky · 2022-06-23T14:40:11.593Z · comments (26)
Covid 6/23/22: Under Five Alive
Zvi · 2022-06-23T14:00:00.958Z · comments (9)
How do states respond to changes in nuclear risk
NathanBarnard · 2022-06-23T12:42:15.855Z · comments (2)
[question] What’s the contingency plan if we get AGI tomorrow?
Yitz (yitz) · 2022-06-23T03:10:27.821Z · answers+comments (23)
[question] What are the best "policy" approaches in worlds where alignment is difficult?
LHA (liam-alexander) · 2022-06-23T01:53:21.574Z · answers+comments (0)
AI Training Should Allow Opt-Out
alyssavance · 2022-06-23T01:33:17.801Z · comments (13)
← previous page (newer posts) · next page (older posts) →