LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

Verification and Transparency
DanielFilan · 2019-08-08T01:50:00.935Z · comments (6)
AI Alignment Open Thread August 2019
habryka (habryka4) · 2019-08-04T22:09:38.431Z · comments (96)
AI Forecasting Resolution Council (Forecasting infrastructure, part 2)
jacobjacob · 2019-08-29T17:35:26.962Z · comments (2)
[question] What authors consistently give accurate pictures of complex topics they discuss?
seez · 2019-08-21T00:09:52.057Z · answers+comments (3)
[meta] [Site Update] Weekly/Monthly/Yearly on All Posts
Raemon · 2019-08-02T00:39:54.461Z · comments (7)
[question] What experiments would demonstrate "upper limits of augmented working memory?"
Raemon · 2019-08-15T22:09:14.492Z · answers+comments (6)
Calibrating With Cards
[deleted] · 2019-08-08T06:44:44.853Z · comments (3)
[link] "Can We Survive Technology" by von Neumann
Ben Pace (Benito) · 2019-08-18T18:58:54.929Z · comments (2)
AI Alignment Writing Day Roundup #1
Ben Pace (Benito) · 2019-08-30T01:26:05.485Z · comments (12)
Distance Functions are Hard
Grue_Slinky · 2019-08-13T17:33:15.296Z · comments (19)
[question] What explanatory power does Kahneman's System 2 possess?
Richard_Ngo (ricraz) · 2019-08-12T15:23:20.197Z · answers+comments (2)
[question] What are the reasons to *not* consider reducing AI-Xrisk the highest priority cause?
David Scott Krueger (formerly: capybaralet) (capybaralet) · 2019-08-20T21:45:12.118Z · answers+comments (27)
When do utility functions constrain?
Hoagy · 2019-08-23T17:19:06.414Z · comments (7)
Help forecast study replication in this social science prediction market
rosiecam · 2019-08-07T18:18:03.102Z · comments (3)
Self-Supervised Learning and AGI Safety
Steven Byrnes (steve2152) · 2019-08-07T14:21:37.739Z · comments (9)
Inspection Paradox as a Driver of Group Separation
shminux · 2019-08-17T21:47:35.812Z · comments (0)
Predicted AI alignment event/meeting calendar
rmoehn · 2019-08-14T07:14:57.233Z · comments (14)
Announcement: Writing Day Today (Thursday)
Ben Pace (Benito) · 2019-08-22T04:48:38.086Z · comments (5)
A Survey of Early Impact Measures
Matthew Barnett (matthew-barnett) · 2019-08-06T01:22:27.421Z · comments (0)
[question] Could we solve this email mess if we all moved to paid emails?
jacobjacob · 2019-08-11T16:31:10.698Z · answers+comments (50)
[AN #62] Are adversarial examples caused by real but imperceptible features?
Rohin Shah (rohinmshah) · 2019-08-22T17:10:01.959Z · comments (10)
GPT-2: 6-Month Follow-Up
[deleted] · 2019-08-21T05:06:52.461Z · comments (1)
[link] "Designing agent incentives to avoid reward tampering", DeepMind
gwern · 2019-08-14T16:57:29.228Z · comments (15)
Don't Pull a Broken Chain
johnswentworth · 2019-08-28T01:21:37.622Z · comments (6)
Algorithmic Similarity
LukasM · 2019-08-23T16:39:47.706Z · comments (10)
[question] Why are the people who could be doing safety research, but aren’t, doing something else?
Adam Scholl (adam_scholl) · 2019-08-29T08:51:33.219Z · answers+comments (19)
Raph Koster on Virtual Worlds vs Games (notes)
Raemon · 2019-08-18T19:01:53.768Z · comments (8)
[question] What is the state of the ego depletion field?
Eli Tyre (elityre) · 2019-08-09T20:30:44.798Z · answers+comments (10)
Negative "eeny meeny miny moe"
jefftk (jkaufman) · 2019-08-20T02:48:41.509Z · comments (6)
Inversion of theorems into definitions when generalizing
riceissa · 2019-08-04T17:44:07.044Z · comments (3)
Project Proposal: Considerations for trading off capabilities and safety impacts of AI research
David Scott Krueger (formerly: capybaralet) (capybaralet) · 2019-08-06T22:22:20.928Z · comments (11)
Goodhart's Curse and Limitations on AI Alignment
Gordon Seidoh Worley (gworley) · 2019-08-19T07:57:01.143Z · comments (18)
Reversible changes: consider a bucket of water
Stuart_Armstrong · 2019-08-26T22:55:23.616Z · comments (18)
A Primer on Matrix Calculus, Part 1: Basic review
Matthew Barnett (matthew-barnett) · 2019-08-12T23:44:37.068Z · comments (3)
Why Gradients Vanish and Explode
Matthew Barnett (matthew-barnett) · 2019-08-09T02:54:44.199Z · comments (9)
[question] Why do humans not have built-in neural i/o channels?
Richard_Ngo (ricraz) · 2019-08-08T13:09:54.072Z · answers+comments (23)
Which of these five AI alignment research projects ideas are no good?
rmoehn · 2019-08-08T07:17:28.959Z · comments (13)
Implications of Quantum Computing for Artificial Intelligence Alignment Research
Jsevillamol · 2019-08-22T10:33:27.502Z · comments (3)
Understanding understanding
mthq · 2019-08-23T18:10:36.796Z · comments (1)
[link] July 2019 gwern.net newsletter
gwern · 2019-08-01T16:19:59.893Z · comments (0)
Emotions are not beliefs
Chris_Leong · 2019-08-07T06:27:49.812Z · comments (2)
[Site Update] Behind the scenes data-layer and caching improvements
habryka (habryka4) · 2019-08-07T00:49:29.721Z · comments (3)
Cartographic Processes
johnswentworth · 2019-08-27T20:02:45.263Z · comments (3)
Practical consequences of impossibility of value learning
Stuart_Armstrong · 2019-08-02T23:06:03.317Z · comments (13)
[question] Do you do weekly or daily reviews? What are they like?
benwr · 2019-08-05T01:23:43.351Z · answers+comments (8)
In defense of Oracle ("Tool") AI research
Steven Byrnes (steve2152) · 2019-08-07T19:14:10.435Z · comments (11)
A Primer on Matrix Calculus, Part 2: Jacobians and other fun
Matthew Barnett (matthew-barnett) · 2019-08-15T01:13:16.070Z · comments (7)
Four Ways An Impact Measure Could Help Alignment
Matthew Barnett (matthew-barnett) · 2019-08-08T00:10:14.304Z · comments (1)
Problems with AI debate
Stuart_Armstrong · 2019-08-26T19:21:40.051Z · comments (3)
[question] Is LW making progress?
zulupineapple · 2019-08-24T00:32:31.600Z · answers+comments (11)
← previous page (newer posts) · next page (older posts) →