LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

n of m ring signatures
DanielFilan · 2023-12-04T20:00:06.580Z · comments (7)
[link] What's new at FAR AI
AdamGleave · 2023-12-04T21:18:03.951Z · comments (0)
[link] 2023 Alignment Research Updates from FAR AI
AdamGleave · 2023-12-04T22:32:19.842Z · comments (0)
Interview with Vanessa Kosoy on the Value of Theoretical Research for AI
WillPetillo · 2023-12-04T22:58:40.005Z · comments (0)
Open Thread – Winter 2023/2024
habryka (habryka4) · 2023-12-04T22:59:49.957Z · comments (160)
Speaking to Congressional staffers about AI risk
Akash (akash-wasil) · 2023-12-04T23:08:52.055Z · comments (23)
[link] Accelerating science through evolvable institutions
jasoncrawford · 2023-12-04T23:21:35.330Z · comments (9)
Bands And Low-stakes Dances
jefftk (jkaufman) · 2023-12-05T03:50:22.076Z · comments (0)
The LessWrong 2022 Review
habryka (habryka4) · 2023-12-05T04:00:00.000Z · comments (43)
Some open-source dictionaries and dictionary learning infrastructure
Sam Marks (samuel-marks) · 2023-12-05T06:05:21.903Z · comments (7)
Analyzing the Historical Rate of Catastrophes
jsteinhardt · 2023-12-05T06:30:01.757Z · comments (0)
Neural uncertainty estimation review article (for alignment)
Charlie Steiner · 2023-12-05T08:01:32.723Z · comments (3)
A Socratic dialogue with my student
lsusr · 2023-12-05T09:31:05.266Z · comments (14)
[link] We're all in this together
Tamsin Leake (carado-1) · 2023-12-05T13:57:46.270Z · comments (65)
On ‘Responsible Scaling Policies’ (RSPs)
Zvi · 2023-12-05T16:10:06.310Z · comments (3)
Deep Forgetting & Unlearning for Safely-Scoped LLMs
scasper · 2023-12-05T16:48:18.177Z · comments (29)
Studying The Alien Mind
Quentin FEUILLADE--MONTIXI (quentin-feuillade-montixi) · 2023-12-05T17:27:28.049Z · comments (10)
[link] In defence of Helen Toner, Adam D'Angelo, and Tasha McCauley (OpenAI post)
mrtreasure · 2023-12-05T18:40:19.740Z · comments (2)
Arguments for/against scheming that focus on the path SGD takes (Section 3 of "Scheming AIs")
Joe Carlsmith (joekc) · 2023-12-05T18:48:12.917Z · comments (0)
Critique-a-Thon of AI Alignment Plans
Iknownothing · 2023-12-05T20:50:07.661Z · comments (3)
[question] How do you feel about LessWrong these days? [Open feedback thread]
jacobjacob · 2023-12-05T20:54:42.317Z · answers+comments (272)
Multinational corporations as optimizers: a case for reaching across the aisle
sudo-nym · 2023-12-06T00:14:35.831Z · comments (10)
ACX Corvallis, OR
kenakofer · 2023-12-06T00:23:25.706Z · comments (0)
Some quick thoughts on "AI is easy to control"
Mikhail Samin (mikhail-samin) · 2023-12-06T00:58:53.681Z · comments (9)
[link] **In defence of Helen Toner, Adam D'Angelo, and Tasha McCauley**
mrtreasure · 2023-12-06T02:02:32.004Z · comments (3)
EA Infrastructure Fund's Plan to Focus on Principles-First EA
Linch · 2023-12-06T03:24:55.844Z · comments (0)
Digital humans vs merge with AI? Same or different?
Nathan Helm-Burger (nathan-helm-burger) · 2023-12-06T04:56:38.261Z · comments (11)
Anthropical Paradoxes are Paradoxes of Probability Theory
Ape in the coat · 2023-12-06T08:16:26.846Z · comments (18)
Minimal Viable Paradise: How do we get The Good Future(TM)?
Nathan Young · 2023-12-06T09:24:09.699Z · comments (0)
[link] Metaculus Launches Chinese AI Chips Tournament, Supporting Institute for AI Policy and Strategy Research
ChristianWilliams · 2023-12-06T11:26:15.790Z · comments (1)
[link] Why Yudkowsky is wrong about "covalently bonded equivalents of biology"
titotal (lombertini) · 2023-12-06T14:09:15.402Z · comments (40)
Bucket Brigade: Likely End-of-Life
jefftk (jkaufman) · 2023-12-06T15:30:06.871Z · comments (1)
Based Beff Jezos and the Accelerationists
Zvi · 2023-12-06T16:00:08.380Z · comments (29)
[link] Google Gemini Announced
Jacob G-W (g-w1) · 2023-12-06T16:14:07.192Z · comments (22)
Proposal for improving the global online discourse through personalised comment ordering on all websites
Roman Leventov · 2023-12-06T18:51:37.645Z · comments (21)
Originality vs. Correctness
alkjash · 2023-12-06T18:51:49.531Z · comments (16)
On Trust
johnswentworth · 2023-12-06T19:19:07.680Z · comments (24)
The counting argument for scheming (Sections 4.1 and 4.2 of "Scheming AIs")
Joe Carlsmith (joekc) · 2023-12-06T19:28:19.393Z · comments (0)
Mathematics As Physics
Nox ML · 2023-12-06T22:27:54.140Z · comments (10)
[question] For fun: How long can you hold your breath?
exanova (yoyo-yuan) · 2023-12-06T23:36:11.320Z · answers+comments (7)
Reflective consistency, randomized decisions, and the dangers of unrealistic thought experiments
Radford Neal · 2023-12-07T03:33:16.149Z · comments (21)
Language Model Memorization, Copyright Law, and Conditional Pretraining Alignment
RogerDearnaley (roger-d-1) · 2023-12-07T06:14:13.816Z · comments (0)
[link] The GiveWiki’s Top Picks in AI Safety for the Giving Season of 2023
Dawn Drescher (Telofy) · 2023-12-07T09:23:05.018Z · comments (10)
Would AIs trapped in the Metaverse pine to enter the real world and would the ramifications cause trouble?
ProfessorFalken · 2023-12-07T10:17:44.732Z · comments (1)
[link] (Report) Evaluating Taiwan's Tactics to Safeguard its Semiconductor Assets Against a Chinese Invasion
Gauraventh (aryangauravyadav) · 2023-12-07T11:50:59.543Z · comments (5)
[question] Is AlphaGo actually a consequentialist utility maximizer?
faul_sname · 2023-12-07T12:41:05.132Z · answers+comments (8)
Random Musings on Theory of Impact for Activation Vectors
Chris_Leong · 2023-12-07T13:07:08.215Z · comments (0)
Gemini 1.0
Zvi · 2023-12-07T14:40:05.243Z · comments (7)
Results from the Turing Seminar hackathon
Charbel-Raphaël (charbel-raphael-segerie) · 2023-12-07T14:50:38.377Z · comments (1)
Simplicity arguments for scheming (Section 4.3 of "Scheming AIs")
Joe Carlsmith (joekc) · 2023-12-07T15:05:54.267Z · comments (1)
← previous page (newer posts) · next page (older posts) →