LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

[link] Linguistic Imperialism in AI: Enforcing Human-Readable Chain-of-Thought
Lukas Petersson (lukas-petersson-1) · 2025-02-21T15:45:00.146Z · comments (0)
Safe Distillation With a Powerful Untrusted AI
Alek Westover (alek-westover) · 2025-02-20T03:14:04.893Z · comments (1)
[link] Teaching AI to reason: this year's most important story
Benjamin_Todd · 2025-02-13T17:40:02.869Z · comments (0)
Information throughput of biological humans and frontier LLMs
benwr · 2025-02-22T07:15:45.457Z · comments (0)
Closed-ended questions aren't as hard as you think
electroswing · 2025-02-19T03:53:11.855Z · comments (0)
[link] Neural Scaling Laws Rooted in the Data Distribution
aribrill (Particleman) · 2025-02-20T21:22:10.306Z · comments (0)
Are current LLMs safe for psychotherapy?
PaperBike · 2025-02-12T19:16:34.452Z · comments (4)
[link] AISN #48: Utility Engineering and EnigmaEval
Corin Katzke (corin-katzke) · 2025-02-18T19:15:16.751Z · comments (0)
A fable on AI x-risk
bgaesop · 2025-02-18T20:15:24.933Z · comments (0)
Permanent properties of things are a self-fulfilling prophecy
YanLyutnev (YanLutnev) · 2025-02-19T00:08:20.776Z · comments (0)
Claude 3.5 Sonnet (New)'s AGI scenario
Nathan Young · 2025-02-17T18:47:04.669Z · comments (2)
Response to the US Govt's Request for Information Concerning Its AI Action Plan
Davey Morse (davey-morse) · 2025-02-14T06:14:08.673Z · comments (0)
[link] Demonstrating specification gaming in reasoning models
Matrice Jacobine · 2025-02-20T19:26:20.563Z · comments (0)
OpenAI’s NSFW policy: user safety, harm reduction, and AI consent
8e9 · 2025-02-13T13:59:22.911Z · comments (3)
Build a Metaculus Forecasting Bot in 30 Minutes: A Practical Guide
ChristianWilliams · 2025-02-22T03:52:14.753Z · comments (0)
Sparse Autoencoder Feature Ablation for Unlearning
aludert · 2025-02-13T19:13:48.388Z · comments (0)
Undesirable Conclusions and Origin Adjustment
Jerdle (daniel-amdurer) · 2025-02-19T18:35:23.732Z · comments (0)
Intrinsic Dimension of Prompts in LLMs
Karthik Viswanathan (vkarthik095) · 2025-02-14T19:02:49.464Z · comments (0)
[link] Probability of AI-Caused Disaster
Alvin Ånestrand (alvin-anestrand) · 2025-02-12T19:40:11.121Z · comments (2)
Artificial Static Place Intelligence: Guaranteed Alignment
ank · 2025-02-15T11:08:50.226Z · comments (2)
arch-anarchist reading list
Peter lawless · 2025-02-16T22:47:00.273Z · comments (1)
[question] Why do we have the NATO logo?
KvmanThinking (avery-liu) · 2025-02-19T22:59:41.755Z · answers+comments (4)
Preference for uncertainty and impact overestimation bias in altruistic systems.
Luck (luck-1) · 2025-02-15T12:27:05.474Z · comments (0)
[link] Modularity and assembly: AI safety via thinking smaller
D Wong (d-nell) · 2025-02-20T00:58:39.714Z · comments (0)
Workshop: Interpretability in LLMs Using Geometric and Statistical Methods
Karthik Viswanathan (vkarthik095) · 2025-02-22T09:39:26.446Z · comments (0)
Biological humans collectively exert at most 400 gigabits/s of control over the world.
benwr · 2025-02-20T23:44:06.509Z · comments (1)
[link] New LLM Scaling Law
wrmedford · 2025-02-19T20:21:17.475Z · comments (0)
Fun, endless art debates v. morally charged art debates that are intrinsically endless
danielechlin · 2025-02-21T04:44:22.712Z · comments (0)
Gradient Anatomy's - Hallucination Robustness in Medical Q&A
DieSab (diego-sabajo) · 2025-02-12T19:16:58.949Z · comments (0)
On Static Space-Like Nature of Intelligence & Superintelligence
ank · 2025-02-22T00:12:36.263Z · comments (0)
[question] Programming Language Early Funding?
J Thomas Moros (J_Thomas_Moros) · 2025-02-16T17:34:06.058Z · answers+comments (5)
Quantifying the Qualitative: Towards a Bayesian Approach to Personal Insight
Pruthvi Kumar (pruthvi-kumar) · 2025-02-15T19:50:42.550Z · comments (0)
Static Place AI Makes AGI Redundant: Multiversal AI Alignment & Rational Utopia
ank · 2025-02-13T22:35:28.300Z · comments (2)
[link] LLMs can teach themselves to better predict the future
Ben Turtel (ben-turtel) · 2025-02-13T01:01:12.175Z · comments (1)
The Newbie's Guide to Navigating AI Futures
keithjmenezes · 2025-02-19T20:37:06.272Z · comments (0)
Places of Loving Grace [Story]
ank · 2025-02-18T23:49:18.580Z · comments (0)
the dumbest theory of everything
lostinwilliamsburg · 2025-02-13T07:57:38.842Z · comments (0)
[link] Sea Change
Charlie Sanders (charlie-sanders) · 2025-02-18T06:03:06.961Z · comments (2)
[link] Humans are Just Self Aware Intelligent Biological Machines
asksathvik · 2025-02-21T01:03:59.950Z · comments (3)
CyberEconomy. The Limits to Growth
Timur Sadekov (timur-sadekov) · 2025-02-16T21:02:34.040Z · comments (0)
Paranoia, Cognitive Biases, and Catastrophic Thought Patterns.
Spiritus Dei (spiritus-dei) · 2025-02-14T00:13:56.300Z · comments (1)
[link] Several Arguments Against the Mathematical Universe Hypothesis
Vittu Perkele · 2025-02-19T22:13:59.425Z · comments (6)
[link] Against Unlimited Genius for Baby-Killers
ggggg · 2025-02-19T20:33:27.188Z · comments (0)
Objective Realism: A Perspective Beyond Human Constructs
Apatheos · 2025-02-14T19:02:33.865Z · comments (1)
The Unearned Privilege We Rarely Discuss: Cognitive Capability
DiegoRojas · 2025-02-18T20:06:45.970Z · comments (7)
[link] On the Rebirth of Aristocracy in the American Regime
shawkisukkar · 2025-02-17T16:18:17.117Z · comments (3)
AI Alignment and the Financial War Against Narcissistic Manipulation
henophilia · 2025-02-19T20:42:10.918Z · comments (2)
Born on Third Base: The Case for Inheriting Nothing and Building Everything
charlieoneill (kingchucky211) · 2025-02-18T00:47:02.298Z · comments (16)
← previous page (newer posts) · next page (older posts) →