LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

larger language models may disappoint you [or, an eternally unfinished draft]
nostalgebraist · 2021-11-26T23:08:56.221Z · comments (31)
Dear Self; we need to talk about ambition
Elizabeth (pktechgirl) · 2023-08-27T23:10:04.720Z · comments (27)
Comment reply: my low-quality thoughts on why CFAR didn't get farther with a "real/efficacious art of rationality"
AnnaSalamon · 2022-06-09T02:12:35.151Z · comments (63)
[link] My PhD thesis: Algorithmic Bayesian Epistemology
Eric Neyman (UnexpectedValues) · 2024-03-16T22:56:59.283Z · comments (14)
UFO Betting: Put Up or Shut Up
RatsWrongAboutUAP · 2023-06-13T04:05:32.652Z · comments (216)
Is Rationalist Self-Improvement Real?
Jacob Falkovich (Jacobian) · 2019-12-09T17:11:03.337Z · comments (78)
So, geez there's a lot of AI content these days
Raemon · 2022-10-06T21:32:20.833Z · comments (140)
Alignment Implications of LLM Successes: a Debate in One Act
Zack_M_Davis · 2023-10-21T15:22:23.053Z · comments (55)
[link] Paul Christiano named as US AI Safety Institute Head of AI Safety
[deleted] · 2024-04-16T16:22:06.937Z · comments (58)
My Model Of EA Burnout
LoganStrohl (BrienneYudkowsky) · 2023-01-25T17:52:42.770Z · comments (50)
Sexual Abuse attitudes might be infohazardous
Pseudonymous Otter · 2022-07-19T18:06:43.956Z · comments (71)
The Plan
johnswentworth · 2021-12-10T23:41:39.417Z · comments (78)
Mental Health and the Alignment Problem: A Compilation of Resources (updated April 2023)
Chris Scammell (chris-scammell) · 2023-05-10T19:04:21.138Z · comments (54)
AI alignment is distinct from its near-term applications
paulfchristiano · 2022-12-13T07:10:04.407Z · comments (21)
The shard theory of human values
Quintin Pope (quintin-pope) · 2022-09-04T04:28:11.752Z · comments (67)
Your Dog is Even Smarter Than You Think
StyleOfDog · 2021-05-01T05:16:09.821Z · comments (108)
What cognitive biases feel like from the inside
chaosmage · 2020-01-03T14:24:22.265Z · comments (32)
Omicron: My Current Model
Zvi · 2021-12-28T17:10:00.629Z · comments (72)
Ngo and Yudkowsky on alignment difficulty
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2021-11-15T20:31:34.135Z · comments (151)
[link] Strong Evidence is Common
Mark Xu (mark-xu) · 2021-03-13T22:04:40.538Z · comments (50)
CFAR Participant Handbook now available to all
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2020-01-03T15:43:44.618Z · comments (40)
Notes from "Don't Shoot the Dog"
juliawise · 2021-04-02T16:34:46.170Z · comments (12)
Coordination as a Scarce Resource
johnswentworth · 2020-01-25T23:32:36.309Z · comments (22)
Thoughts on the impact of RLHF research
paulfchristiano · 2023-01-25T17:23:16.402Z · comments (102)
My Assessment of the Chinese AI Safety Community
Lao Mein (derpherpize) · 2023-04-25T04:21:19.274Z · comments (94)
Laziness death spirals
PatrickDFarley · 2024-09-19T15:58:30.252Z · comments (36)
Visible Thoughts Project and Bounty Announcement
So8res · 2021-11-30T00:19:08.408Z · comments (106)
On AutoGPT
Zvi · 2023-04-13T12:30:01.059Z · comments (47)
[link] I hired 5 people to sit behind me and make me productive for a month
Simon Berens (sberens) · 2023-02-05T01:19:39.182Z · comments (83)
The Best Lay Argument is not a Simple English Yud Essay
J Bostock (Jemist) · 2024-09-10T17:34:28.422Z · comments (15)
Yes, It's Subjective, But Why All The Crabs?
johnswentworth · 2023-07-28T19:35:36.741Z · comments (15)
My views on “doom”
paulfchristiano · 2023-04-27T17:50:01.415Z · comments (37)
You Don't Exist, Duncan
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2023-02-02T08:37:01.049Z · comments (107)
The LessWrong Team is now Lightcone Infrastructure, come work with us!
habryka (habryka4) · 2021-10-01T01:20:33.411Z · comments (71)
The ground of optimization
Alex Flint (alexflint) · 2020-06-20T00:38:15.521Z · comments (80)
Truthseeking is the ground in which other principles grow
Elizabeth (pktechgirl) · 2024-05-27T01:09:20.796Z · comments (16)
[link] DeepMind: Generally capable agents emerge from open-ended play
Daniel Kokotajlo (daniel-kokotajlo) · 2021-07-27T14:19:13.782Z · comments (53)
New Scaling Laws for Large Language Models
1a3orn · 2022-04-01T20:41:17.665Z · comments (22)
Working With Monsters
johnswentworth · 2021-07-20T15:23:20.762Z · comments (54)
The Feeling of Idea Scarcity
johnswentworth · 2022-12-31T17:34:04.306Z · comments (22)
Ilya Sutskever and Jan Leike resign from OpenAI [updated]
Zach Stein-Perlman · 2024-05-15T00:45:02.436Z · comments (95)
Principles for the AGI Race
William_S · 2024-08-30T14:29:41.074Z · comments (13)
the case for CoT unfaithfulness is overstated
nostalgebraist · 2024-09-29T22:07:54.053Z · comments (41)
Deep Deceptiveness
So8res · 2023-03-21T02:51:52.794Z · comments (60)
Lessons On How To Get Things Right On The First Try
johnswentworth · 2023-06-19T23:58:09.605Z · comments (57)
Orienting to 3 year AGI timelines
Nikola Jurkovic (nikolaisalreadytaken) · 2024-12-22T01:15:11.401Z · comments (47)
Munk AI debate: confusions and possible cruxes
Steven Byrnes (steve2152) · 2023-06-27T14:18:47.694Z · comments (21)
Another (outer) alignment failure story
paulfchristiano · 2021-04-07T20:12:32.043Z · comments (38)
How "Discovering Latent Knowledge in Language Models Without Supervision" Fits Into a Broader Alignment Scheme
Collin (collin-burns) · 2022-12-15T18:22:40.109Z · comments (39)
RadVac Commercial Antibody Test Results
johnswentworth · 2021-02-26T18:04:09.171Z · comments (30)
← previous page (newer posts) · next page (older posts) →