LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

The Sinews of Sudan’s Latest War
Tim Liptrot (rockthecasbah) · 2023-08-04T18:17:27.860Z · comments (12)
[link] Read More Books but Pretend to Read Even More
Arjun Panickssery (arjun-panickssery) · 2023-08-05T00:07:48.671Z · comments (12)
[link] Announcing Squiggle Hub
ozziegooen · 2023-08-05T01:00:17.739Z · comments (4)
[question] What are the best published papers from outside the alignment community that are relevant to Agent Foundations?
Stephen Fowler (LosPolloFowler) · 2023-08-05T03:02:33.003Z · answers+comments (4)
Meet Hyperion on Sunday Aug 6?
duck_master · 2023-08-05T04:36:02.462Z · comments (0)
ACX Paris Meetup - August 11 2023
PoignardAzur · 2023-08-05T09:44:05.717Z · comments (0)
A Naive Proposal for Constructing Interpretable AI
Chris_Leong · 2023-08-05T10:32:05.446Z · comments (6)
[Linkpost] Applicability of scaling laws to vision encoding models
Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2023-08-05T11:10:35.599Z · comments (2)
video games > IQ tests
bhauth · 2023-08-05T13:27:54.697Z · comments (35)
[link] Stomach Ulcers and Dental Cavities
Metacelsus · 2023-08-05T14:08:15.263Z · comments (7)
[link] Join AISafety.info's Writing & Editing Hackathon (Aug 25-28) (Prizes to be won!)
smallsilo (monstrologies) · 2023-08-05T14:08:19.639Z · comments (3)
AISafety.info's Writing & Editing Hackathon
smallsilo (monstrologies) · 2023-08-05T17:14:45.292Z · comments (0)
Seattle Astral Codex Ten Monthly Social
a7x · 2023-08-05T17:55:25.884Z · comments (0)
[link] Ground-Truth Label Imbalance Impairs the Performance of Contrast-Consistent Search (and Other Contrast-Pair-Based Unsupervised Methods)
Tom Angsten (tom-angsten) · 2023-08-05T17:55:46.569Z · comments (2)
Summary of Improving Global Decision Making (around AI)
Will_Pearson · 2023-08-05T18:46:44.268Z · comments (0)
how 2 tell if ur input is out of distribution given only model weights
dkirmani · 2023-08-05T22:45:20.250Z · comments (10)
Aligning my web server with devops practices: part 2 (security)
VipulNaik · 2023-08-06T01:30:35.005Z · comments (0)
Exploring the Multiverse of Large Language Models
franky · 2023-08-06T02:38:02.784Z · comments (0)
The Benevolent Ruler’s Handbook (Part 1): The Policy Problem
FCCC · 2023-08-06T03:46:31.594Z · comments (3)
Safety-First Agents/Architectures Are a Promising Path to Safe AGI
Brendon_Wong · 2023-08-06T08:02:30.072Z · comments (2)
[question] On being in a bad place and too stubborn to leave.
TeaTieAndHat (Augustin Portier) · 2023-08-06T11:45:49.771Z · answers+comments (14)
[link] Model-Based Policy Analysis under Deep Uncertainty
Max Reddel (max-reddel) · 2023-08-06T14:07:36.079Z · comments (1)
[link] Rebooting AI Governance: An AI-Driven Approach to AI Governance
Max Reddel (max-reddel) · 2023-08-06T14:19:50.180Z · comments (1)
Reducing the risk of catastrophically misaligned AI by avoiding the Singleton scenario: the Manyton Variant
GravitasGradient (Bll) · 2023-08-06T14:24:04.774Z · comments (0)
[Linkpost] Will AI avoid exploitation?
cdkg · 2023-08-06T14:28:29.166Z · comments (1)
[link] ‘We’re changing the clouds.’ An unforeseen test of geoengineering is fueling record ocean warmth
Annapurna (jorge-velez) · 2023-08-06T20:58:51.838Z · comments (6)
Computational Thread Art
CallumMcDougall (TheMcDouglas) · 2023-08-06T21:42:30.306Z · comments (2)
[link] Yann LeCun on AGI and AI Safety
Chris_Leong · 2023-08-06T21:56:52.644Z · comments (13)
Problems with Robin Hanson's Quillette Article On AI
DaemonicSigil · 2023-08-06T22:13:43.654Z · comments (33)
Drinks at a bar
yakimoff · 2023-08-07T02:52:19.388Z · comments (0)
The second act: Beginning epistemic rigor at 30
hiAndrewQuinn (hiandrewquinn) · 2023-08-07T09:34:20.923Z · comments (0)
[link] Overview of how AI might exacerbate long-running catastrophic risks
Hauke Hillebrandt (hauke-hillebrandt) · 2023-08-07T11:53:29.171Z · comments (0)
Strengthening the Argument for Intrinsic AI Safety: The S-Curves Perspective
avturchin · 2023-08-07T13:13:42.635Z · comments (0)
Monthly Roundup #9: August 2023
Zvi · 2023-08-07T13:20:03.522Z · comments (25)
[link] What I've been reading, July–August 2023
jasoncrawford · 2023-08-07T14:22:57.046Z · comments (0)
[link] Announcing the Clearer Thinking micro-grants program for 2023
spencerg · 2023-08-07T15:21:28.191Z · comments (1)
Optimisation Measures: Desiderata, Impossibility, Proposals
mattmacdermott · 2023-08-07T15:52:17.624Z · comments (9)
[question] Should I test myself for microplastics?
Augs · 2023-08-07T17:31:41.656Z · answers+comments (2)
[link] Growing Bonsai Networks with RNNs
ameo (ameobea) · 2023-08-07T17:34:15.713Z · comments (5)
Feedbackloop-first Rationality
Raemon · 2023-08-07T17:58:56.349Z · comments (65)
[link] An interactive introduction to grokking and mechanistic interpretability
Adam Pearce (adam-pearce) · 2023-08-07T19:09:19.422Z · comments (3)
[question] Tips for reducing thinking branching factor
Simon Berens (sberens) · 2023-08-07T20:21:43.298Z · answers+comments (6)
A plea for more funding shortfall transparency
porby · 2023-08-07T21:33:11.912Z · comments (4)
[question] How do I find all the items on LW that I've *favorited* or upvoted?
Alex K. Chen (parrot) (alex-k-chen) · 2023-08-07T23:51:05.711Z · answers+comments (3)
Perpetually Declining Population?
jefftk (jkaufman) · 2023-08-08T01:30:00.897Z · comments (29)
Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research
evhub · 2023-08-08T01:30:10.847Z · comments (26)
Notice your everything
metachirality · 2023-08-08T02:38:39.974Z · comments (1)
4 types of AGI selection, and how to constrain them
Remmelt (remmelt-ellen) · 2023-08-08T10:02:53.921Z · comments (3)
My Trial Period as an Independent Alignment Researcher
Bart Bussmann (Stuckwork) · 2023-08-08T14:16:35.122Z · comments (1)
[question] Beginner's question about RLHF
[deleted] · 2023-08-08T15:48:24.118Z · answers+comments (3)
← previous page (newer posts) · next page (older posts) →