LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

State of Generally Available Self-Driving
jefftk (jkaufman) · 2023-08-22T18:50:01.166Z · comments (6)
Steven Wolfram on AI Alignment
Bill Benzon (bill-benzon) · 2023-08-20T19:49:28.953Z · comments (15)
LLMs are (mostly) not helped by filler tokens
Kshitij Sachan (kshitij-sachan) · 2023-08-10T00:48:50.510Z · comments (35)
[link] AI Forecasting: Two Years In
jsteinhardt · 2023-08-19T23:40:04.302Z · comments (15)
“Dirty concepts” in AI alignment discourses, and some guesses for how to deal with them
Nora_Ammann · 2023-08-20T09:13:34.225Z · comments (4)
AI Regulation May Be More Important Than AI Alignment For Existential Safety
otto.barten (otto-barten) · 2023-08-24T11:41:54.690Z · comments (39)
Aumann-agreement is common
tailcalled · 2023-08-26T20:22:03.738Z · comments (31)
Ideas for improving epistemics in AI safety outreach
mic (michael-chen) · 2023-08-21T19:55:45.654Z · comments (6)
"Is There Anything That's Worth More"
Zack_M_Davis · 2023-08-02T03:28:16.116Z · comments (6)
[link] What Does a Marginal Grant at LTFF Look Like? Funding Priorities and Grantmaking Thresholds at the Long-Term Future Fund
Linch · 2023-08-11T03:59:51.757Z · comments (0)
A short calculation about a Twitter poll
Ege Erdil (ege-erdil) · 2023-08-14T19:48:53.018Z · comments (64)
When Omnipotence is Not Enough
lsusr · 2023-08-25T19:50:51.038Z · comments (2)
[link] DIY Deliberate Practice
lynettebye · 2023-08-21T12:22:10.284Z · comments (4)
Private notes on LW?
Raemon · 2023-08-04T17:35:37.917Z · comments (33)
[link] ‘We’re changing the clouds.’ An unforeseen test of geoengineering is fueling record ocean warmth
Annapurna (jorge-velez) · 2023-08-06T20:58:51.838Z · comments (6)
AI #25: Inflection Point
Zvi · 2023-08-17T14:40:06.940Z · comments (9)
AI #23: Fundamental Problems with RLHF
Zvi · 2023-08-03T12:50:11.852Z · comments (9)
Diet Experiment Preregistration: Long-term water fasting + seed oil removal
lc · 2023-08-23T22:08:49.058Z · comments (17)
[link] Stomach Ulcers and Dental Cavities
Metacelsus · 2023-08-05T14:08:15.263Z · comments (7)
[link] Will AI kill everyone? Here's what the godfathers of AI have to say [RA video]
Writer · 2023-08-19T17:29:04.227Z · comments (8)
Open Call for Research Assistants in Developmental Interpretability
Jesse Hoogland (jhoogland) · 2023-08-30T09:02:59.781Z · comments (11)
If we had known the atmosphere would ignite
Jeffs · 2023-08-16T20:28:51.166Z · comments (49)
Reflections on "Making the Atomic Bomb"
boazbarak · 2023-08-17T02:48:19.933Z · comments (7)
Why Is No One Trying To Align Profit Incentives With Alignment Research?
Prometheus · 2023-08-23T13:16:41.927Z · comments (11)
The lost millennium
Ege Erdil (ege-erdil) · 2023-08-24T03:48:40.035Z · comments (14)
AI #26: Fine Tuning Time
Zvi · 2023-08-24T15:30:06.626Z · comments (6)
Barbieheimer: Across the Dead Reckoning
Zvi · 2023-08-01T13:00:05.700Z · comments (17)
AI #24: Week of the Podcast
Zvi · 2023-08-10T15:00:04.438Z · comments (5)
Perpetually Declining Population?
jefftk (jkaufman) · 2023-08-08T01:30:00.897Z · comments (29)
Assessment of intelligence agency functionality is difficult yet important
trevor (TrevorWiesinger) · 2023-08-24T01:42:20.931Z · comments (5)
Efficiency and resource use scaling parity
Ege Erdil (ege-erdil) · 2023-08-21T00:18:01.243Z · comments (0)
how 2 tell if ur input is out of distribution given only model weights
dkirmani · 2023-08-05T22:45:20.250Z · comments (10)
[link] Announcing Squiggle Hub
ozziegooen · 2023-08-05T01:00:17.739Z · comments (4)
Chess as a case study in hidden capabilities in ChatGPT
AdamYedidia (babybeluga) · 2023-08-19T06:35:03.459Z · comments (32)
Understanding and visualizing sycophancy datasets
Nina Rimsky (NinaR) · 2023-08-16T05:34:06.899Z · comments (0)
A Model-based Approach to AI Existential Risk
Sammy Martin (SDM) · 2023-08-25T10:32:16.817Z · comments (9)
[link] Manifund: What we're funding (weeks 2-4)
Austin Chen (austin-chen) · 2023-08-04T16:00:33.227Z · comments (2)
Is Chinese total factor productivity lower today than it was in 1956?
Ege Erdil (ege-erdil) · 2023-08-18T22:33:50.560Z · comments (0)
The Sinews of Sudan’s Latest War
Tim Liptrot (rockthecasbah) · 2023-08-04T18:17:27.860Z · comments (12)
[question] Which possible AI systems are relatively safe?
Zach Stein-Perlman · 2023-08-21T17:00:27.582Z · answers+comments (20)
Monthly Roundup #9: August 2023
Zvi · 2023-08-07T13:20:03.522Z · comments (25)
Autonomous replication and adaptation: an attempt at a concrete danger threshold
Hjalmar_Wijk · 2023-08-17T01:31:10.554Z · comments (0)
[Linkpost] Personal and Psychological Dimensions of AI Researchers Confronting AI Catastrophic Risks
Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2023-08-12T22:02:09.895Z · comments (0)
AI Deception: A Survey of Examples, Risks, and Potential Solutions
Simon Goldstein (simon-goldstein) · 2023-08-29T01:29:50.916Z · comments (3)
[link] Walk while you talk: don't balk at "no chalk"
dkl9 · 2023-08-22T21:27:47.257Z · comments (9)
[link] marine cloud brightening
bhauth · 2023-08-09T02:50:56.639Z · comments (14)
AGI is easier than robotaxis
Daniel Kokotajlo (daniel-kokotajlo) · 2023-08-13T17:00:29.901Z · comments (30)
[link] Implications of evidential cooperation in large worlds
Lukas Finnveden (Lanrian) · 2023-08-23T00:43:45.232Z · comments (4)
[link] Seth Explains Consciousness
Jacob Falkovich (Jacobian) · 2023-08-22T18:06:42.653Z · comments (125)
Learning as you play: anthropic shadow in deadly games
dr_s · 2023-08-12T07:34:42.261Z · comments (28)
← previous page (newer posts) · next page (older posts) →