LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

How to solve deception and still fail.
Charlie Steiner · 2023-10-04T19:56:56.254Z · comments (7)
Fraternal Birth Order Effect and the Maternal Immune Hypothesis
Bucky · 2023-10-04T21:18:08.753Z · comments (0)
Underspecified Probabilities: A Thought Experiment
lunatic_at_large · 2023-10-04T22:25:07.458Z · comments (4)
On my AI Fable, and the importance of de re, de dicto, and de se reference for AI alignment
PhilGoetz · 2023-10-05T00:50:43.012Z · comments (4)
How to Get Rationalist Feedback
NicholasKross · 2023-10-05T02:03:10.766Z · comments (0)
Response to Quintin Pope's Evolution Provides No Evidence For the Sharp Left Turn
Zvi · 2023-10-05T11:39:02.393Z · comments (29)
Making progress on the ``what alignment target should be aimed at?'' question, is urgent
ThomasCederborg · 2023-10-05T12:55:06.736Z · comments (0)
Can the House Legislate?
jefftk (jkaufman) · 2023-10-05T13:40:06.649Z · comments (6)
AI #32: Lie Detector
Zvi · 2023-10-05T13:50:05.030Z · comments (19)
This anime storyboard doesn't exist: a graphic novel written and illustrated by GPT4
RomanS · 2023-10-05T14:01:30.141Z · comments (7)
Twin Cities ACX Meetup October 2023
Timothy M. (timothy-bond) · 2023-10-05T16:29:29.761Z · comments (2)
Censorship in LLMs is here to stay because it mirrors how our own intelligence is structured
mnvr · 2023-10-05T17:37:48.425Z · comments (0)
Translations Should Invert
abramdemski · 2023-10-05T17:44:23.262Z · comments (19)
Evaluating the historical value misspecification argument
Matthew Barnett (matthew-barnett) · 2023-10-05T18:34:15.695Z · comments (140)
A well-defined history in measurable factor spaces
Matthias G. Mayer (matthias-georg-mayer) · 2023-10-05T18:36:05.037Z · comments (0)
Ideation and Trajectory Modelling in Language Models
NickyP (Nicky) · 2023-10-05T19:21:07.990Z · comments (2)
[link] Towards Monosemanticity: Decomposing Language Models With Dictionary Learning
Zac Hatfield-Dodds (zac-hatfield-dodds) · 2023-10-05T21:01:39.767Z · comments (21)
Impacts of AI on the housing markets
PottedRosePetal · 2023-10-05T21:24:03.045Z · comments (0)
Stampy's AI Safety Info soft launch
steven0461 · 2023-10-05T22:13:04.632Z · comments (9)
[link] Provably Safe AI
PeterMcCluskey · 2023-10-05T22:18:26.013Z · comments (15)
[link] The Long-Term Future Fund is looking for a full-time fund chair
Linch · 2023-10-05T22:18:53.720Z · comments (0)
A personal explanation of ELK concept and task.
Zeyu Qin (zeyu-qin) · 2023-10-06T03:55:44.895Z · comments (0)
Super-Exponential versus Exponential Growth in Compute Price-Performance
moridinamael · 2023-10-06T16:23:56.714Z · comments (24)
Linking Alt Accounts
jefftk (jkaufman) · 2023-10-06T17:00:09.802Z · comments (33)
Don't Dismiss Simple Alignment Approaches
Chris_Leong · 2023-10-07T00:35:26.789Z · comments (9)
Announcing Dialogues
Ben Pace (Benito) · 2023-10-07T02:57:39.005Z · comments (52)
On the Relationship Between Variability and the Evolutionary Outcomes of Systems in Nature
Artyom Shaposhnikov (artyom-shaposhnikov) · 2023-10-07T03:06:54.184Z · comments (0)
[question] Current State of Probabilistic Logic
lunatic_at_large · 2023-10-07T05:06:22.955Z · answers+comments (2)
Related Discussion from Thomas Kwa's MIRI Research Experience
Raemon · 2023-10-07T06:25:00.994Z · comments (140)
Contra Nora Belrose on Orthogonality Thesis Being Trivial
tailcalled · 2023-10-07T11:47:02.401Z · comments (21)
Fixing Insider Threats in the AI Supply Chain
Madhav Malhotra (madhav-malhotra) · 2023-10-07T13:19:17.320Z · comments (2)
[link] Every Mention of EA in "Going Infinite"
KirstenH · 2023-10-07T14:42:32.217Z · comments (0)
Griffin Island
jefftk (jkaufman) · 2023-10-07T18:40:06.642Z · comments (3)
Sam Altman's sister, Annie Altman, claims Sam has severely abused her
pl5015 · 2023-10-07T21:06:49.396Z · comments (105)
Arguments for utilitarianism are impossibility arguments under unbounded prospects
MichaelStJules · 2023-10-07T21:08:59.645Z · comments (7)
A thought about the constraints of debtlessness in online communities
mako yass (MakoYass) · 2023-10-07T21:26:44.480Z · comments (23)
Comparing Anthropic's Dictionary Learning to Ours
Robert_AIZI · 2023-10-07T23:30:32.402Z · comments (8)
[link] The Gradient – The Artificiality of Alignment
mic (michael-chen) · 2023-10-08T04:06:40.074Z · comments (1)
Perspective Based Reasoning Could Absolve CDT
dadadarren · 2023-10-08T11:22:49.458Z · comments (5)
[link] Bird-eye view visualization of LLM activations
Sergii (sergey-kharagorgiev) · 2023-10-08T12:12:25.593Z · comments (2)
[link] Linkpost: Are Emergent Abilities in Large Language Models just In-Context Learning?
Erich_Grunewald · 2023-10-08T12:14:46.825Z · comments (6)
Time is homogeneous sequentially-composable determination
TsviBT · 2023-10-08T14:58:15.913Z · comments (0)
FAQ: What the heck is goal agnosticism?
porby · 2023-10-08T19:11:50.269Z · comments (36)
[link] "The Heart of Gaming is the Power Fantasy", and Cohabitive Games
Raemon · 2023-10-08T21:02:33.526Z · comments (49)
AI Alignment Breakthroughs this week (10/08/23)
Logan Zoellner (logan-zoellner) · 2023-10-08T23:30:54.924Z · comments (14)
Paper: Identifying the Risks of LM Agents with an LM-Emulated Sandbox - University of Toronto 2023 - Benchmark consisting of 36 high-stakes tools and 144 test cases!
Singularian2501 (maik-zywitza) · 2023-10-09T00:00:19.859Z · comments (0)
Leveraging Bayes' Theorem to Supercharge Memory Techniques
disoha · 2023-10-09T03:34:04.958Z · comments (1)
We don't understand what happened with culture enough
Jan_Kulveit · 2023-10-09T09:54:20.096Z · comments (21)
Knowledge Base 2: The structure and the method of building
iwis · 2023-10-09T11:53:14.689Z · comments (4)
Knowledge Base 3: Shopping advisor and other uses of knowledge base about products
iwis · 2023-10-09T11:53:35.065Z · comments (0)
← previous page (newer posts) · next page (older posts) →