LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Aversion Factoring
CFAR!Duncan (CFAR 2017) · 2022-07-07T16:09:11.392Z · comments (1)
A Pattern Language For Rationality
Vaniver · 2022-07-05T19:08:49.783Z · comments (14)
Which values are stable under ontology shifts?
Richard_Ngo (ricraz) · 2022-07-23T02:40:04.344Z · comments (48)
[link] A time-invariant version of Laplace's rule
Jsevillamol · 2022-07-15T19:28:15.877Z · comments (13)
Principles of Privacy for Alignment Research
johnswentworth · 2022-07-27T19:53:28.209Z · comments (30)
[link] NeurIPS ML Safety Workshop 2022
Dan H (dan-hendrycks) · 2022-07-26T15:28:52.441Z · comments (2)
Cognitive Risks of Adolescent Binge Drinking
Elizabeth (pktechgirl) · 2022-07-20T21:10:01.513Z · comments (12)
Avoid the abbreviation "FLOPs" – use "FLOP" or "FLOP/s" instead
Daniel_Eth · 2022-07-10T10:44:38.046Z · comments (13)
Abstracting The Hardness of Alignment: Unbounded Atomic Optimization
adamShimi · 2022-07-29T18:59:49.460Z · comments (3)
My vision of a good future, part I
Jeffrey Ladish (jeff-ladish) · 2022-07-06T01:23:01.074Z · comments (18)
Curating "The Epistemic Sequences" (list v.0.1)
Andrew_Critch · 2022-07-23T22:17:08.544Z · comments (12)
Applications are open for CFAR workshops in Prague this fall!
John Steidley (JohnSteidley) · 2022-07-19T18:29:19.172Z · comments (3)
Taste & Shaping
CFAR!Duncan (CFAR 2017) · 2022-07-10T05:50:14.416Z · comments (1)
What's next for instrumental rationality?
Andrew_Critch · 2022-07-23T22:55:06.185Z · comments (7)
Response to Blake Richards: AGI, generality, alignment, & loss functions
Steven Byrnes (steve2152) · 2022-07-12T13:56:00.885Z · comments (9)
Introducing the Fund for Alignment Research (We're Hiring!)
AdamGleave · 2022-07-06T02:07:47.965Z · comments (0)
[link] My Most Likely Reason to Die Young is AI X-Risk
AISafetyIsNotLongtermist · 2022-07-04T17:08:27.209Z · comments (24)
Conditioning Generative Models for Alignment
Jozdien · 2022-07-18T07:11:46.369Z · comments (8)
When Giving People Money Doesn’t Help
Zvi · 2022-07-07T13:00:00.879Z · comments (12)
A Bias Against Altruism
Lone Pine (conor-sullivan) · 2022-07-23T20:44:59.964Z · comments (30)
[link] Deep learning curriculum for large language model alignment
Jacob_Hilton · 2022-07-13T21:58:33.452Z · comments (3)
The Reader's Guide to Optimal Monetary Policy
Ege Erdil (ege-erdil) · 2022-07-25T15:10:51.010Z · comments (10)
Double Crux
CFAR!Duncan (CFAR 2017) · 2022-07-24T06:34:15.305Z · comments (9)
Deception?! I ain’t got time for that!
Paul Colognese (paul-colognese) · 2022-07-18T00:06:15.274Z · comments (5)
[AN #172] Sorry for the long hiatus!
Rohin Shah (rohinmshah) · 2022-07-05T06:20:03.943Z · comments (0)
Don't take the organizational chart literally
lc · 2022-07-21T00:56:28.561Z · comments (21)
Making decisions using multiple worldviews
Richard_Ngo (ricraz) · 2022-07-13T19:15:02.621Z · comments (10)
[link] Acceptability Verification: A Research Agenda
David Udell · 2022-07-12T20:11:34.986Z · comments (0)
Race Along Rashomon Ridge
Stephen Fowler (LosPolloFowler) · 2022-07-07T03:20:59.701Z · comments (15)
Outer vs inner misalignment: three framings
Richard_Ngo (ricraz) · 2022-07-06T19:46:50.902Z · comments (5)
Report from a civilizational observer on Earth
owencb · 2022-07-09T17:26:09.223Z · comments (12)
Comfort Zone Exploration
CFAR!Duncan (CFAR 2017) · 2022-07-15T21:18:14.033Z · comments (2)
Announcing the AI Safety Field Building Hub, a new effort to provide AISFB projects, mentorship, and funding
Vael Gates · 2022-07-28T21:29:52.424Z · comments (3)
Goal Alignment Is Robust To the Sharp Left Turn
Thane Ruthenis · 2022-07-13T20:23:58.962Z · comments (16)
Potato diet: A post mortem and an answer to SMTM's article
Épiphanie Gédéon (joy_void_joy) · 2022-07-14T23:18:26.691Z · comments (34)
The Alignment Problem
lsusr · 2022-07-11T03:03:03.271Z · comments (18)
Babysitting as Parenting Trial?
jefftk (jkaufman) · 2022-07-07T13:20:04.129Z · comments (19)
[link] The Most Important Century: The Animation
Writer · 2022-07-24T20:58:55.869Z · comments (2)
Eavesdropping on Aliens: A Data Decoding Challenge
anonymousaisafety · 2022-07-24T04:35:40.880Z · comments (9)
Tarnished Guy who Puts a Num on it
Jacob Falkovich (Jacobian) · 2022-07-06T18:05:59.168Z · comments (11)
Artificial Sandwiching: When can we test scalable alignment protocols without humans?
Sam Bowman (sbowman) · 2022-07-13T21:14:08.145Z · comments (6)
Safety considerations for online generative modeling
Sam Marks (samuel-marks) · 2022-07-07T18:31:19.316Z · comments (9)
The curious case of Pretty Good human inner/outer alignment
PavleMiha · 2022-07-05T19:04:49.434Z · comments (45)
Systemization
CFAR!Duncan (CFAR 2017) · 2022-07-11T18:39:04.750Z · comments (5)
[link] [Linkpost] Existential Risk Analysis in Empirical Research Papers
Dan H (dan-hendrycks) · 2022-07-02T00:09:49.399Z · comments (0)
Bucket Errors
CFAR!Duncan (CFAR 2017) · 2022-07-29T18:50:48.549Z · comments (7)
[link] QNR Prospects
PeterMcCluskey · 2022-07-16T02:03:37.258Z · comments (3)
Covid 7/14/22: BA.2.75 Plus Tax
Zvi · 2022-07-14T14:40:00.587Z · comments (9)
Mosaic and Palimpsests: Two Shapes of Research
adamShimi · 2022-07-12T09:05:28.984Z · comments (3)
Predicting Parental Emotional Changes?
jefftk (jkaufman) · 2022-07-06T13:50:04.387Z · comments (11)
← previous page (newer posts) · next page (older posts) →