LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

[link] AISN #29: Progress on the EU AI Act Plus, the NY Times sues OpenAI for Copyright Infringement, and Congressional Questions about Research Standards in AI Safety
aogara (Aidan O'Gara) · 2024-01-04T16:09:31.336Z · comments (0)
Some Vacation Photos
johnswentworth · 2024-01-04T17:15:01.187Z · comments (0)
Deep atheism and AI risk
Joe Carlsmith (joekc) · 2024-01-04T18:58:47.745Z · comments (22)
[link] Cellular reprogramming, pneumatic launch systems, and terraforming Mars: Some things I learned about at Foresight Vision Weekend
jasoncrawford · 2024-01-04T19:33:57.887Z · comments (0)
The Gears of Argmax
StrivingForLegibility · 2024-01-04T23:30:30.339Z · comments (0)
Safety Data Sheets for Optimization Processes
StrivingForLegibility · 2024-01-04T23:30:36.510Z · comments (1)
Best-Responding Is Not Always the Best Response
StrivingForLegibility · 2024-01-04T23:30:48.400Z · comments (0)
Using Threats to Achieve Socially Optimal Outcomes
StrivingForLegibility · 2024-01-04T23:30:54.615Z · comments (0)
Hello
S Benfield (steven-benfield) · 2024-01-04T23:35:05.621Z · comments (0)
[link] Project ideas: Governance during explosive technological growth
Lukas Finnveden (Lanrian) · 2024-01-04T23:51:56.407Z · comments (0)
MIRI 2024 Mission and Strategy Update
Malo (malo) · 2024-01-05T00:20:54.169Z · comments (44)
Does AI care about reality or just its own perception?
RedFishBlueFish (RedStateBlueState) · 2024-01-05T04:05:11.167Z · comments (8)
If I ran the zoo
Optimization Process · 2024-01-05T05:14:57.631Z · comments (0)
Striking Implications for Learning Theory, Interpretability — and Safety?
RogerDearnaley (roger-d-1) · 2024-01-05T08:46:58.915Z · comments (4)
Predictive model agents are sort of corrigible
Raymond D · 2024-01-05T14:05:03.037Z · comments (6)
[link] Forecast your 2024 with Fatebook
Sage Future (aaron-ho-1) · 2024-01-05T14:07:55.743Z · comments (0)
AI Impacts Survey: December 2023 Edition
Zvi · 2024-01-05T14:40:06.156Z · comments (6)
Catching AIs red-handed
ryan_greenblatt · 2024-01-05T17:43:10.948Z · comments (18)
[question] What technical topics could help with boundaries/membranes?
Chipmonk · 2024-01-05T18:14:58.795Z · answers+comments (25)
[link] The Hippie Rabbit Hole -Nuggets of Gold in Rivers of Bullshit
Jonathan Moregård (JonathanMoregard) · 2024-01-05T18:27:01.769Z · comments (20)
Technology path dependence and evaluating expertise
bhauth · 2024-01-05T19:21:23.302Z · comments (2)
[link] AI Impacts 2023 Expert Survey on Progress in AI
habryka (habryka4) · 2024-01-05T19:42:17.226Z · comments (1)
The Next ChatGPT Moment: AI Avatars
kolmplex (luke-man) · 2024-01-05T20:14:10.074Z · comments (10)
[link] Almost everyone I’ve met would be well-served thinking more about what to focus on
Henrik Karlsson (henrik-karlsson) · 2024-01-05T21:01:27.861Z · comments (8)
[link] Benchmark Study #1: MMLU (Pile, MCQ)
Bruce W. Lee (bruce-lee) · 2024-01-05T21:35:37.999Z · comments (0)
[link] Project ideas: Epistemics
Lukas Finnveden (Lanrian) · 2024-01-05T23:41:23.721Z · comments (4)
[link] Benchmark Study #2: TruthfulQA (Task, MCQ)
Bruce W. Lee (bruce-lee) · 2024-01-06T02:39:39.895Z · comments (2)
Survey of 2,778 AI authors: six parts in pictures
KatjaGrace · 2024-01-06T04:43:34.590Z · comments (1)
Are we inside a black hole?
Jay · 2024-01-06T13:30:51.451Z · comments (5)
Book review: Trick or treatment (2008)
Fleece Minutia · 2024-01-06T15:40:49.953Z · comments (0)
A Land Tax For Britain
A.H. (AlfredHarwood) · 2024-01-06T15:52:14.942Z · comments (9)
Lack of Spider-Man is evidence against the simulation hypothesis
RamblinDash · 2024-01-06T18:17:20.641Z · comments (22)
A Challenge to Effective Altruism's Premises
False Name (False Name, Esq.) · 2024-01-06T18:46:23.715Z · comments (3)
AI Risk and the US Presidential Candidates
Zane · 2024-01-06T20:18:04.945Z · comments (22)
The Sequences on YouTube
Neil (neil-warren) · 2024-01-07T01:44:39.663Z · comments (9)
[link] Defending against hypothetical moon life during Apollo 11
eukaryote · 2024-01-07T04:49:42.628Z · comments (9)
[link] Benchmark Study #3: HellaSwag (Task, MCQ)
Bruce W. Lee (bruce-lee) · 2024-01-07T04:59:21.347Z · comments (4)
[link] Towards AI Safety Infrastructure: Talk & Outline
Paul Bricman (paulbricman) · 2024-01-07T09:31:12.217Z · comments (0)
[link] Bayesians Commit the Gambler's Fallacy
Kevin Dorst · 2024-01-07T12:54:59.939Z · comments (28)
Deceptive AI ≠ Deceptively-aligned AI
Steven Byrnes (steve2152) · 2024-01-07T16:55:13.761Z · comments (19)
Benchmark Study #4: AI2 Reasoning Challenge (Task(s), MCQ)
Bruce W. Lee (bruce-lee) · 2024-01-07T17:13:00.209Z · comments (0)
[link] Project ideas: Sentience and rights of digital minds
Lukas Finnveden (Lanrian) · 2024-01-07T17:34:58.942Z · comments (0)
(Partial) failure in replicating deceptive alignment experiment
claudia.biancotti · 2024-01-07T17:56:36.748Z · comments (0)
We shouldn't fear superintelligence because it already exists
Spencer Chubb (spencer-chubb) · 2024-01-07T17:59:55.297Z · comments (14)
[link] A model of research skill
L Rudolf L (LRudL) · 2024-01-08T00:13:12.755Z · comments (6)
Utility is relative
CrimsonChin · 2024-01-08T02:31:44.000Z · comments (4)
Sledding Among Hazards
jefftk (jkaufman) · 2024-01-08T03:30:08.463Z · comments (5)
Why There Is Hope For An Alignment Solution
Darklight · 2024-01-08T06:58:32.820Z · comments (0)
Reflections on my first year of AI safety research
Jay Bailey · 2024-01-08T07:49:08.147Z · comments (3)
There is no sharp boundary between deontology and consequentialism
quetzal_rainbow · 2024-01-08T11:01:47.828Z · comments (2)
← previous page (newer posts) · next page (older posts) →