LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Path dependence in ML inductive biases
Vivek Hebbar (Vivek) · 2022-09-10T01:38:22.885Z · comments (13)
Quintin's alignment papers roundup - week 2
Quintin Pope (quintin-pope) · 2022-09-19T13:41:27.104Z · comments (2)
Where I currently disagree with Ryan Greenblatt’s version of the ELK approach
So8res · 2022-09-29T21:18:44.402Z · comments (7)
Book review: “The Heart of the Brain: The Hypothalamus and Its Hormones”
Steven Byrnes (steve2152) · 2022-09-27T13:20:51.434Z · comments (3)
A game of mattering
KatjaGrace · 2022-09-23T02:30:15.714Z · comments (7)
LOVE in a simbox is all you need
jacob_cannell · 2022-09-28T18:25:31.283Z · comments (72)
[Closed] Prize and fast track to alignment research at ALTER
Vanessa Kosoy (vanessa-kosoy) · 2022-09-17T16:58:24.839Z · comments (6)
[link] Self-Control Secrets of the Puritan Masters
David Hugh-Jones (david-hugh-jones) · 2022-09-26T09:04:56.895Z · comments (3)
Private alignment research sharing and coordination
porby · 2022-09-04T00:01:22.337Z · comments (13)
Gradient Hacker Design Principles From Biology
johnswentworth · 2022-09-01T19:03:16.836Z · comments (13)
[link] Argument against 20% GDP growth from AI within 10 years [Linkpost]
aogara (Aidan O'Gara) · 2022-09-12T04:08:03.901Z · comments (21)
Clarifying the Agent-Like Structure Problem
johnswentworth · 2022-09-29T21:28:08.813Z · comments (15)
Fake qualities of mind
Kaj_Sotala · 2022-09-22T16:40:05.085Z · comments (2)
[link] Review of Examine.com’s vitamin write-ups
Elizabeth (pktechgirl) · 2022-09-26T23:40:06.344Z · comments (1)
QAPR 3: interpretability-guided training of neural nets
Quintin Pope (quintin-pope) · 2022-09-28T16:02:10.732Z · comments (2)
Replacement for PONR concept
Daniel Kokotajlo (daniel-kokotajlo) · 2022-09-02T00:09:45.698Z · comments (6)
Two reasons we might be closer to solving alignment than it seems
KatWoods (ea247) · 2022-09-24T20:00:08.442Z · comments (9)
Levelling Up in AI Safety Research Engineering
Gabe M (gabe-mukobi) · 2022-09-02T04:59:42.699Z · comments (9)
Why deceptive alignment matters for AGI safety
Marius Hobbhahn (marius-hobbhahn) · 2022-09-15T13:38:53.219Z · comments (13)
Infra-Exercises, Part 1
Diffractor · 2022-09-01T05:06:59.373Z · comments (10)
Deep Q-Networks Explained
Jay Bailey · 2022-09-13T12:01:08.033Z · comments (6)
[link] Why was progress so slow in the past?
jasoncrawford · 2022-09-01T20:26:06.163Z · comments (31)
Methodological Therapy: An Agenda For Tackling Research Bottlenecks
adamShimi · 2022-09-22T18:41:03.346Z · comments (6)
We may be able to see sharp left turns coming
Ethan Perez (ethan-perez) · 2022-09-03T02:55:45.168Z · comments (29)
When would AGIs engage in conflict?
JesseClifton · 2022-09-14T19:38:22.478Z · comments (5)
Triangle Opportunity
Alex Beyman (alexbeyman) · 2022-09-26T20:42:30.393Z · comments (10)
[link] First we shape our social graph; then it shapes us
Henrik Karlsson (henrik-karlsson) · 2022-09-07T15:50:08.281Z · comments (6)
[link] ACT-1: Transformer for Actions
Daniel Kokotajlo (daniel-kokotajlo) · 2022-09-14T19:09:39.725Z · comments (4)
When does technical work to reduce AGI conflict make a difference?: Introduction
JesseClifton · 2022-09-14T19:38:00.760Z · comments (3)
Many therapy schools work with inner multiplicity (not just IFS)
David Althaus (wallowinmaya) · 2022-09-17T10:27:41.350Z · comments (15)
EA & LW Forums Weekly Summary (28 Aug - 3 Sep 22’)
Zoe Williams (GreyArea) · 2022-09-06T11:06:25.230Z · comments (2)
Some notes on solving hard problems
Joe Rocca (joseph-rocca) · 2022-09-19T12:58:45.306Z · comments (8)
My Thoughts on the ML Safety Course
zeshen · 2022-09-27T13:15:03.000Z · comments (3)
Coordinate-Free Interpretability Theory
johnswentworth · 2022-09-14T23:33:49.910Z · comments (16)
[link] Dan Luu on Futurist Predictions
RobertM (T3t) · 2022-09-14T03:01:27.275Z · comments (9)
Soft skills for meetups
mingyuan · 2022-09-27T17:26:12.406Z · comments (3)
A Library and Tutorial for Factored Cognition with Language Models
stuhlmueller · 2022-09-28T18:15:10.800Z · comments (0)
Prize idea: Transmit MIRI and Eliezer's worldviews
elifland · 2022-09-19T21:21:13.156Z · comments (18)
[link] ethics and anthropics of homomorphically encrypted computations
Tamsin Leake (carado-1) · 2022-09-09T10:49:08.316Z · comments (49)
Covid 9/29/22: The Jones Act Waver
Zvi · 2022-09-29T18:20:02.103Z · comments (10)
[An email with a bunch of links I sent an experienced ML researcher interested in learning about Alignment / x-safety.]
David Scott Krueger (formerly: capybaralet) (capybaralet) · 2022-09-08T22:28:54.534Z · comments (1)
[link] Scraping training data for your mind
Henrik Karlsson (henrik-karlsson) · 2022-09-21T16:27:48.499Z · comments (4)
Brief Notes on Transformers
Adam Jermyn (adam-jermyn) · 2022-09-26T14:46:23.637Z · comments (3)
Pretending not to Notice
jefftk (jkaufman) · 2022-09-19T02:30:05.079Z · comments (12)
[link] Estimating the Current and Future Number of AI Safety Researchers
Stephen McAleese (stephen-mcaleese) · 2022-09-28T21:11:33.703Z · comments (14)
AI Risk Intro 1: Advanced AI Might Be Very Bad
CallumMcDougall (TheMcDouglas) · 2022-09-11T10:57:12.093Z · comments (13)
AI Safety field-building projects I'd like to see
Akash (akash-wasil) · 2022-09-11T23:43:32.031Z · comments (7)
Samotsvety's AI risk forecasts
elifland · 2022-09-09T04:01:18.958Z · comments (0)
Searching for Modularity in Large Language Models
NickyP (Nicky) · 2022-09-08T02:25:31.711Z · comments (3)
[link] Summaries: Alignment Fundamentals Curriculum
Leon Lang (leon-lang) · 2022-09-18T13:08:05.335Z · comments (3)
← previous page (newer posts) · next page (older posts) →