LessWrong 2.0 Reader

View: New · Old · Top

next page (older posts) →

[link] [Linkpost] Faith and Fate: Limits of Transformers on Compositionality
Joe Kwon · 2023-06-16T15:04:59.828Z · comments (None)
The ones who endure
Richard_Ngo (ricraz) · 2023-06-16T14:40:09.623Z · comments (None)
[link] Conjecture: A standing offer for public debates on AI
Andrea_Miotti (AndreaM) · 2023-06-16T14:33:43.273Z · comments (None)
[link] Explaining "Taking features out of superposition with sparse autoencoders"
Robert_AIZI · 2023-06-16T13:59:27.950Z · comments (None)
[question] How not to write the Cookbook of Doom?
brunoparga · 2023-06-16T13:37:02.328Z · answers+comments (None)
Scaffolded LLMs: Less Obvious Concerns
Stephen Fowler (LosPolloFowler) · 2023-06-16T10:39:58.835Z · comments (2)
Motivation in AI
nickasaf · 2023-06-16T09:50:24.254Z · comments (1)
Distilling Singular Learning Theory
Liam Carroll (liam-carroll) · 2023-06-16T09:50:14.474Z · comments (None)
DSLT 1. The RLCT Measures the Effective Dimension of Singular Models
Liam Carroll (liam-carroll) · 2023-06-16T09:50:10.113Z · comments (None)
[Linkpost] Mapping Brains with Language Models: A Survey
Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2023-06-16T09:49:23.043Z · comments (None)
Rational Animations is looking for an AI Safety scriptwriter, a lead community manager, and other roles.
Writer · 2023-06-16T09:41:14.343Z · comments (None)
Dreaming of Utility
Mariven · 2023-06-16T05:10:12.099Z · comments (None)
[question] Does anyone's full-time job include reading and understanding all the most-promising formal AI alignment work?
NicholasKross · 2023-06-16T02:24:31.048Z · answers+comments (2)
Leveling Up Or Leveling Off? Understanding The Science Behind Skill Plateaus
lynettebye · 2023-06-16T00:18:04.378Z · comments (4)
human intelligence may be alignment-limited
bhauth · 2023-06-15T22:32:14.685Z · comments (3)
[link] Developing a technology with safety in mind: Lessons from the Wright Brothers
jasoncrawford · 2023-06-15T21:08:55.828Z · comments (4)
AXRP Episode 22 - Shard Theory with Quintin Pope
DanielFilan · 2023-06-15T19:00:01.340Z · comments (1)
Can we accelerate human progress? Moderated Conversation in NYC
Jannik Schg (jannik-schilling) · 2023-06-15T17:33:25.356Z · comments (None)
Group Prioritarianism: Why AI Should Not Replace Humanity [draft]
fsh · 2023-06-15T17:33:18.333Z · comments (None)
Press the happiness button!
Spiarrow · 2023-06-15T17:30:26.992Z · comments (3)
[link] [Linkpost] World first as UK hosts inaugural AUKUS AI and autonomy trial
NinaR · 2023-06-15T14:13:10.209Z · comments (None)
Philosophical Cyborg (Part 2)...or, The Good Successor
ukc10014 · 2023-06-15T13:52:28.321Z · comments (None)
AI #16: AI in the UK
Zvi · 2023-06-15T13:20:03.939Z · comments (11)
[link] I still think it's very unlikely we're observing alien aircraft
dynomight · 2023-06-15T13:01:27.734Z · comments (45)
[link] Aligned Objectives Prize Competition
Prometheus · 2023-06-15T12:42:21.682Z · comments (None)
A more effective Elevator Pitch for AI risk
Iknownothing · 2023-06-15T12:39:03.363Z · comments (None)
Why "AI alignment" would better be renamed into "Artificial Intention research"
chaosmage · 2023-06-15T10:32:26.094Z · comments (12)
[link] Matt Taibbi's COVID reporting
ChristianKl · 2023-06-15T09:49:54.272Z · comments (23)
Looking Back On Ads
jefftk (jkaufman) · 2023-06-15T02:10:04.076Z · comments (9)
Why libertarians are advocating for regulation on AI
RobertM (T3t) · 2023-06-14T20:59:58.225Z · comments (13)
Instrumental Convergence? [Draft]
J. Dmitri Gallow (j-dmitri-gallow) · 2023-06-14T20:21:41.485Z · comments (11)
On the Apple Vision Pro
Zvi · 2023-06-14T17:50:08.957Z · comments (12)
[link] Progress links and tweets, 2023-06-14
jasoncrawford · 2023-06-14T16:30:12.346Z · comments (1)
Philosophical Cyborg (Part 1)
ukc10014 · 2023-06-14T16:20:40.317Z · comments (4)
[link] Is the confirmation bias really a bias?
Lionel (lionel) · 2023-06-14T14:06:06.315Z · comments (5)
Apply Now - NA East Organizer Retreat
Willa (Eh_Yo_Lexa) · 2023-06-14T13:39:06.152Z · comments (None)
Lightcone Infrastructure/LessWrong is looking for funding
habryka (habryka4) · 2023-06-14T04:45:53.425Z · comments (24)
[link] Anthropic | Charting a Path to AI Accountability
Gabriel Mukobi (gabe-mukobi) · 2023-06-14T04:43:33.563Z · comments (1)
Demystifying Born's rule
Christopher King (christopher-king) · 2023-06-14T03:16:20.941Z · comments (26)
My guess for why I was wrong about US housing
romeostevensit · 2023-06-14T00:37:04.162Z · comments (11)
Notes from the Bank of England Talk by Giovanni Dosi on Agent-based Modeling for Macroeconomics
PixelatedPenguin · 2023-06-13T22:25:00.769Z · comments (None)
Introducing The Long Game Project: Improving Decision-Making Through Tabletop Exercises and Simulated Experience
Dan Stuart (DrDanEpstein) · 2023-06-13T21:45:02.265Z · comments (None)
Intelligence allocation from a Mean Field Game Theory perspective
Marv K · 2023-06-13T19:52:14.260Z · comments (None)
[link] Multiple stages of fallacy - justifications and non-justifications for the multiple stage fallacy
AronT (Aron Thomas) · 2023-06-13T17:37:20.762Z · comments (2)
TryContra Events
jefftk (jkaufman) · 2023-06-13T17:30:01.597Z · comments (None)
MetaAI: less is less for alignment.
Cleo Nardo (strawberry calm) · 2023-06-13T14:08:45.209Z · comments (10)
The Dial of Progress
Zvi · 2023-06-13T13:40:06.354Z · comments (96)
Virtual AI Safety Unconference (VAISU)
Linda Linsefors · 2023-06-13T09:56:22.542Z · comments (None)
Seattle ACX Meetup - Summer 2023
Optimization Process · 2023-06-13T05:14:38.790Z · comments (None)
TASRA: A Taxonomy and Analysis of Societal-Scale Risks from AI
Andrew_Critch · 2023-06-13T05:04:46.756Z · comments (1)
next page (older posts) →