LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

[link] [SEE NEW EDITS] No, *You* Need to Write Clearer
Nicholas / Heather Kross (NicholasKross) · 2023-04-29T05:04:01.559Z · comments (65)
Comment reply: my low-quality thoughts on why CFAR didn't get farther with a "real/efficacious art of rationality"
AnnaSalamon · 2022-06-09T02:12:35.151Z · comments (63)
larger language models may disappoint you [or, an eternally unfinished draft]
nostalgebraist · 2021-11-26T23:08:56.221Z · comments (31)
The Plan
johnswentworth · 2021-12-10T23:41:39.417Z · comments (78)
Omicron Variant Post #1: We’re F***ed, It’s Never Over
Zvi · 2021-11-26T19:00:00.988Z · comments (95)
Safetywashing
Adam Scholl (adam_scholl) · 2022-07-01T11:56:33.495Z · comments (20)
UFO Betting: Put Up or Shut Up
RatsWrongAboutUAP · 2023-06-13T04:05:32.652Z · comments (216)
[link] My PhD thesis: Algorithmic Bayesian Epistemology
Eric Neyman (UnexpectedValues) · 2024-03-16T22:56:59.283Z · comments (14)
So, geez there's a lot of AI content these days
Raemon · 2022-10-06T21:32:20.833Z · comments (140)
Alignment Implications of LLM Successes: a Debate in One Act
Zack_M_Davis · 2023-10-21T15:22:23.053Z · comments (55)
Orienting to 3 year AGI timelines
Nikola Jurkovic (nikolaisalreadytaken) · 2024-12-22T01:15:11.401Z · comments (49)
Sexual Abuse attitudes might be infohazardous
Pseudonymous Otter · 2022-07-19T18:06:43.956Z · comments (72)
[link] Paul Christiano named as US AI Safety Institute Head of AI Safety
Joel Burget (joel-burget) · 2024-04-16T16:22:06.937Z · comments (58)
My Model Of EA Burnout
LoganStrohl (BrienneYudkowsky) · 2023-01-25T17:52:42.770Z · comments (50)
Mental Health and the Alignment Problem: A Compilation of Resources (updated April 2023)
Chris Scammell (chris-scammell) · 2023-05-10T19:04:21.138Z · comments (54)
AI alignment is distinct from its near-term applications
paulfchristiano · 2022-12-13T07:10:04.407Z · comments (21)
The shard theory of human values
Quintin Pope (quintin-pope) · 2022-09-04T04:28:11.752Z · comments (67)
Your Dog is Even Smarter Than You Think
StyleOfDog · 2021-05-01T05:16:09.821Z · comments (108)
What cognitive biases feel like from the inside
chaosmage · 2020-01-03T14:24:22.265Z · comments (32)
CFAR Participant Handbook now available to all
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2020-01-03T15:43:44.618Z · comments (40)
Ngo and Yudkowsky on alignment difficulty
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2021-11-15T20:31:34.135Z · comments (151)
[link] Strong Evidence is Common
Mark Xu (mark-xu) · 2021-03-13T22:04:40.538Z · comments (50)
Omicron: My Current Model
Zvi · 2021-12-28T17:10:00.629Z · comments (72)
Notes from "Don't Shoot the Dog"
juliawise · 2021-04-02T16:34:46.170Z · comments (12)
Coordination as a Scarce Resource
johnswentworth · 2020-01-25T23:32:36.309Z · comments (22)
Thoughts on the impact of RLHF research
paulfchristiano · 2023-01-25T17:23:16.402Z · comments (102)
Laziness death spirals
PatrickDFarley · 2024-09-19T15:58:30.252Z · comments (36)
My Assessment of the Chinese AI Safety Community
Lao Mein (derpherpize) · 2023-04-25T04:21:19.274Z · comments (94)
My views on “doom”
paulfchristiano · 2023-04-27T17:50:01.415Z · comments (37)
Visible Thoughts Project and Bounty Announcement
So8res · 2021-11-30T00:19:08.408Z · comments (106)
The Best Lay Argument is not a Simple English Yud Essay
J Bostock (Jemist) · 2024-09-10T17:34:28.422Z · comments (15)
The ground of optimization
Alex Flint (alexflint) · 2020-06-20T00:38:15.521Z · comments (80)
Yes, It's Subjective, But Why All The Crabs?
johnswentworth · 2023-07-28T19:35:36.741Z · comments (15)
On AutoGPT
Zvi · 2023-04-13T12:30:01.059Z · comments (47)
[link] I hired 5 people to sit behind me and make me productive for a month
Simon Berens (sberens) · 2023-02-05T01:19:39.182Z · comments (83)
[link] DeepMind: Generally capable agents emerge from open-ended play
Daniel Kokotajlo (daniel-kokotajlo) · 2021-07-27T14:19:13.782Z · comments (53)
Truthseeking is the ground in which other principles grow
Elizabeth (pktechgirl) · 2024-05-27T01:09:20.796Z · comments (16)
the case for CoT unfaithfulness is overstated
nostalgebraist · 2024-09-29T22:07:54.053Z · comments (42)
You Don't Exist, Duncan
Duncan Sabien (Deactivated) (Duncan_Sabien) · 2023-02-02T08:37:01.049Z · comments (107)
The LessWrong Team is now Lightcone Infrastructure, come work with us!
habryka (habryka4) · 2021-10-01T01:20:33.411Z · comments (71)
The Feeling of Idea Scarcity
johnswentworth · 2022-12-31T17:34:04.306Z · comments (22)
Working With Monsters
johnswentworth · 2021-07-20T15:23:20.762Z · comments (54)
New Scaling Laws for Large Language Models
1a3orn · 2022-04-01T20:41:17.665Z · comments (22)
Ilya Sutskever and Jan Leike resign from OpenAI [updated]
Zach Stein-Perlman · 2024-05-15T00:45:02.436Z · comments (95)
Principles for the AGI Race
William_S · 2024-08-30T14:29:41.074Z · comments (13)
Another (outer) alignment failure story
paulfchristiano · 2021-04-07T20:12:32.043Z · comments (38)
Deep Deceptiveness
So8res · 2023-03-21T02:51:52.794Z · comments (60)
Lessons On How To Get Things Right On The First Try
johnswentworth · 2023-06-19T23:58:09.605Z · comments (57)
Munk AI debate: confusions and possible cruxes
Steven Byrnes (steve2152) · 2023-06-27T14:18:47.694Z · comments (21)
How "Discovering Latent Knowledge in Language Models Without Supervision" Fits Into a Broader Alignment Scheme
Collin (collin-burns) · 2022-12-15T18:22:40.109Z · comments (39)
← previous page (newer posts) · next page (older posts) →