LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

Formalizing the "AI x-risk is unlikely because it is ridiculous" argument
Christopher King (christopher-king) · 2023-05-03T18:56:25.834Z · comments (17)
AGI rising: why we are in a new era of acute risk and increasing public awareness, and what to do now
[deleted] · 2023-05-03T20:26:19.479Z · comments (12)
[question] What constraints does deep learning place on alignment plans?
Garrett Baker (D0TheMath) · 2023-05-03T20:40:16.007Z · answers+comments (0)
«Boundaries/Membranes» and AI safety compilation
Chipmonk · 2023-05-03T21:41:19.124Z · comments (17)
[link] Alignment Research @ EleutherAI
Curtis Huebner · 2023-05-03T22:45:25.662Z · comments (1)
Prizes for matrix completion problems
paulfchristiano · 2023-05-03T23:30:08.069Z · comments (51)
Moving VPS Again
jefftk (jkaufman) · 2023-05-04T00:30:04.556Z · comments (2)
How MATS addresses “mass movement building” concerns
Ryan Kidd (ryankidd44) · 2023-05-04T00:55:26.913Z · comments (9)
[New] Rejected Content Section
Ruby · 2023-05-04T01:43:19.547Z · comments (21)
Understanding why illusionism does not deny the existence of qualia
Mergimio H. Doefevmil · 2023-05-04T02:13:57.597Z · comments (17)
Opinion merging for AI control
David Johnston (david-johnston) · 2023-05-04T02:43:51.196Z · comments (0)
Has the Symbol Grounding Problem just gone away?
RussellThor · 2023-05-04T07:46:09.444Z · comments (3)
We don’t need AGI for an amazing future
Karl von Wendt · 2023-05-04T12:10:59.536Z · comments (32)
Advice for interacting with busy people
Severin T. Seehrich (sts) · 2023-05-04T13:31:06.709Z · comments (4)
AI #10: Code Interpreter and Geoff Hinton
Zvi · 2023-05-04T14:00:00.366Z · comments (7)
[link] [Crosspost] AI X-risk in the News: How Effective are Recent Media Items and How is Awareness Changing? Our New Survey Results.
otto.barten (otto-barten) · 2023-05-04T14:09:00.134Z · comments (0)
Clarifying and predicting AGI
Richard_Ngo (ricraz) · 2023-05-04T15:55:26.283Z · comments (42)
[link] [Linkpost]Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Curtis Huebner · 2023-05-04T17:16:33.335Z · comments (1)
Trying to measure AI deception capabilities using temporary simulation fine-tuning
alenoach (glerzing) · 2023-05-04T17:59:28.149Z · comments (0)
[link] Google "We Have No Moat, And Neither Does OpenAI"
Chris_Leong · 2023-05-04T18:23:09.121Z · comments (28)
AI risk/reward: A simple model
Nathan Young · 2023-05-04T19:25:25.738Z · comments (0)
Recursive Middle Manager Hell: AI Edition
VojtaKovarik · 2023-05-04T20:08:17.583Z · comments (11)
[link] Who regulates the regulators? We need to go beyond the review-and-approval paradigm
jasoncrawford · 2023-05-04T22:11:17.465Z · comments (29)
Robin Hanson and I talk about AI risk
KatjaGrace · 2023-05-04T22:20:08.448Z · comments (8)
A brief collection of Hinton's recent comments on AGI risk
Kaj_Sotala · 2023-05-04T23:31:06.157Z · comments (9)
An Update On The Campaign For AI Safety Dot Org
yanni kyriacos (yanni) · 2023-05-05T00:21:56.648Z · comments (2)
How to get good at programming
Ulisse Mini (ulisse-mini) · 2023-05-05T01:14:29.401Z · comments (3)
Transcript of a presentation on catastrophic risks from AI
RobertM (T3t) · 2023-05-05T01:38:17.948Z · comments (0)
[question] What is it like to be a compatibilist?
tslarm · 2023-05-05T02:56:45.084Z · answers+comments (72)
Monthly Shorts 4/23
Celer · 2023-05-05T07:20:06.430Z · comments (1)
Investigating Emergent Goal-Like Behavior in Large Language Models using Experimental Economics
phelps-sg · 2023-05-05T11:15:13.390Z · comments (1)
[question] Why not use active SETI to prevent AI Doom?
RomanS · 2023-05-05T14:41:40.661Z · answers+comments (13)
What can we learn from Bayes about reasoning?
jasoncrawford · 2023-05-05T15:52:35.980Z · comments (11)
Kingfisher Live CD Process
jefftk (jkaufman) · 2023-05-05T17:00:04.651Z · comments (0)
[link] Regulate or Compete? The China Factor in U.S. AI Policy (NAIR #2)
charles_m (charles-martinet) · 2023-05-05T17:43:42.417Z · comments (1)
[link] Linkpost for Accursed Farms Discussion / debate with AI expert Eliezer Yudkowsky
gilch · 2023-05-05T18:20:20.004Z · comments (2)
CHAT Diplomacy: LLMs and National Security
JohnBuridan · 2023-05-05T19:45:27.507Z · comments (6)
[link] What Jason has been reading, May 2023: “Protopia,” complex systems, Daedalus vs. Icarus, and more
jasoncrawford · 2023-05-05T19:54:16.377Z · comments (2)
A smart enough LLM might be deadly simply if you run it for long enough
Mikhail Samin (mikhail-samin) · 2023-05-05T20:49:31.416Z · comments (16)
[link] Orthogonal's Formal-Goal Alignment theory of change
Tamsin Leake (carado-1) · 2023-05-05T22:36:14.883Z · comments (12)
Sleeping Beauty – the Death Hypothesis
Guillaume Charrier (guillaume-charrier) · 2023-05-05T23:32:20.278Z · comments (6)
Explaining “Hell is Game Theory Folk Theorems”
electroswing · 2023-05-05T23:33:20.977Z · comments (21)
Do you work at an AI lab? Please quit
Nik Samoylov (nik-samoylov) · 2023-05-05T23:41:31.560Z · comments (9)
Naturalist Collection
LoganStrohl (BrienneYudkowsky) · 2023-05-06T00:37:42.774Z · comments (7)
You don't need to be a genius to be in AI safety research
Claire Short (claire-short) · 2023-05-06T02:32:02.164Z · comments (1)
My preferred framings for reward misspecification and goal misgeneralisation
Yi-Yang (yiyang) · 2023-05-06T04:48:49.257Z · comments (1)
Alignment as Function Fitting
A.H. (AlfredHarwood) · 2023-05-06T11:38:04.245Z · comments (0)
Estimating Norovirus Prevalence
jefftk (jkaufman) · 2023-05-06T11:40:04.588Z · comments (0)
The Broader Fossil Fuel Community
Jeffrey Heninger (jeffrey-heninger) · 2023-05-06T14:49:38.633Z · comments (1)
Is "red" for GPT-4 the same as "red" for you?
Yusuke Hayashi (hayashiyus) · 2023-05-06T17:55:20.691Z · comments (6)
← previous page (newer posts) · next page (older posts) →