LessWrong 2.0 Reader

View: New · Old · Top

Restrict date range: Today · This week · This month · Last three months · This year · All time

← previous page (newer posts) · next page (older posts) →

On AI Detectors Regarding College Applications
Kaustubh Kislay (kaustubh-kislay) · 2024-11-27T20:25:48.151Z · comments (1)
Germany-wide ACX Meetup
Fernand0 · 2024-11-17T10:08:54.584Z · comments (0)
[question] Is OpenAI net negative for AI Safety?
Lysandre Terrisse · 2024-11-02T16:18:02.859Z · answers+comments (0)
Some Comments on Recent AI Safety Developments
testingthewaters · 2024-11-09T16:44:58.936Z · comments (0)
[question] Noticing the World
EvolutionByDesign (bioluminescent-darkness) · 2024-11-04T16:41:44.696Z · answers+comments (1)
A better “Statement on AI Risk?”
Knight Lee (Max Lee) · 2024-11-25T04:50:29.399Z · comments (4)
What are Emotions?
Myles H (zarsou9) · 2024-11-15T04:20:27.388Z · comments (13)
Visualizing small Attention-only Transformers
WCargo (Wcargo) · 2024-11-19T09:37:42.213Z · comments (0)
Effects of Non-Uniform Sparsity on Superposition in Toy Models
Shreyans Jain (shreyans-jain) · 2024-11-14T16:59:43.234Z · comments (3)
Understanding Emergence in Large Language Models
egek92 · 2024-11-29T19:42:43.790Z · comments (1)
Hope to live or fear to die?
Knight Lee (Max Lee) · 2024-11-27T10:42:37.070Z · comments (0)
Distributed espionage
margetmagenta · 2024-11-04T19:43:33.316Z · comments (0)
The boat
RomanS · 2024-11-22T12:56:45.050Z · comments (0)
[link] When the Scientific Method Doesn't Really Help...
casualphysicsenjoyer (hatta_afiq) · 2024-11-27T19:52:30.023Z · comments (1)
AI Training Opt-Outs Reinforce Global Power Asymmetries
kushagra (kushagra-tiwari) · 2024-11-30T22:08:06.426Z · comments (0)
Should you increase AI alignment funding, or increase AI regulation?
Knight Lee (Max Lee) · 2024-11-26T09:17:01.809Z · comments (1)
Sexual Selection as a Mesa-Optimizer
Lorec · 2024-11-29T23:34:45.739Z · comments (0)
[link] Decorated pedestrian tunnels
dkl9 · 2024-11-24T22:16:03.794Z · comments (3)
Reducing x-risk might be actively harmful
MountainPath · 2024-11-18T14:25:07.127Z · comments (5)
[link] Both-Sidesism—When Fair & Balanced Goes Wrong
James Stephen Brown (james-brown) · 2024-11-02T03:04:03.820Z · comments (15)
Beyond Gaussian: Language Model Representations and Distributions
Matt Levinson · 2024-11-24T01:53:38.156Z · comments (1)
Workshop Report: Why current benchmarks approaches are not sufficient for safety?
Tom DAVID (tom-david) · 2024-11-26T17:20:47.453Z · comments (1)
[link] Higher Order Signs, Hallucination and Schizophrenia
Nicolas Villarreal (nicolas-villarreal) · 2024-11-02T16:33:10.574Z · comments (0)
LDT (and everything else) can be irrational
Christopher King (christopher-king) · 2024-11-06T04:05:36.932Z · comments (6)
notes on prioritizing tasks & cognition-threads
Emrik (Emrik North) · 2024-11-26T00:28:03.400Z · comments (1)
Antonym Heads Predict Semantic Opposites in Language Models
Jake Ward (jake-ward) · 2024-11-15T15:32:14.102Z · comments (0)
[link] Sparks of Consciousness
Charlie Sanders (charlie-sanders) · 2024-11-13T04:58:27.222Z · comments (0)
Which AI Safety Benchmark Do We Need Most in 2025?
Loïc Cabannes (loic-cabannes) · 2024-11-17T23:50:56.337Z · comments (2)
The 'Road Not Taken' in the Multiverse
Jonah Wilberg (jrwilb@googlemail.com) · 2024-11-29T19:01:51.775Z · comments (0)
A Meritocracy of Taste
Daniele De Nuntiis (daniele-de-nuntiis) · 2024-11-28T09:10:10.598Z · comments (10)
ESTIMAT: A Fermi-Based Framework for Cognitive State Optimization
Gabriel Brito (gabriel-brito) · 2024-12-01T15:13:55.699Z · comments (3)
Don't want Goodhart? — Specify the variables more
YanLyutnev (YanLutnev) · 2024-11-21T22:43:48.362Z · comments (2)
Root node of my posts
AtillaYasar (atillayasar) · 2024-11-19T20:09:02.973Z · comments (0)
[question] Poll: what’s your impression of altruism?
David Gross (David_Gross) · 2024-11-09T20:28:15.418Z · answers+comments (4)
[question] Have we seen any "ReLU instead of sigmoid-type improvements" recently
KvmanThinking (avery-liu) · 2024-11-23T03:51:52.984Z · answers+comments (4)
aspirational leadership
dhruvmethi · 2024-11-20T16:07:43.507Z · comments (0)
How to make evals for the AISI evals bounty
TheManxLoiner · 2024-12-03T10:44:45.700Z · comments (0)
MIT FutureTech are hiring ‍a Product and Data Visualization Designer
peterslattery · 2024-11-13T14:48:06.167Z · comments (0)
Breaking beliefs about saving the world
Oxidize · 2024-11-15T00:46:03.693Z · comments (3)
Agenda Manipulation
Pazzaz · 2024-11-09T14:13:33.729Z · comments (0)
Launching Applications for the Global AI Safety Fellowship 2025!
Aditya_SK (team-ai-safety) · 2024-11-30T14:02:16.537Z · comments (2)
Launching a 5-day Intro to Transformative AI course
bluedotimpact · 2024-11-22T17:45:05.304Z · comments (0)
Optimizing Problem-Solving Strategies Through Prediction Markets
patrik-cihal · 2024-11-22T19:58:32.930Z · comments (0)
Modeling AI-driven occupational change over the next 10 years and beyond
2120eth · 2024-11-12T04:58:26.741Z · comments (0)
Truth Terminal: A reconstruction of events
crvr.fr (crdevio) · 2024-11-17T23:51:21.279Z · comments (1)
[question] A Coordination Cookbook?
azergante · 2024-11-10T23:20:34.843Z · answers+comments (0)
[question] Will Orion/Gemini 2/Llama-4 outperform o1
LuigiPagani (luigipagani) · 2024-11-18T21:15:55.953Z · answers+comments (3)
AI alignment via civilizational cognitive updates
AtillaYasar (atillayasar) · 2024-11-10T09:33:35.023Z · comments (10)
Jakarta ACX December 2024 Meetup
Aud (aud) · 2024-11-19T15:01:31.101Z · comments (0)
[link] Paradigm Shifts—change everything... except almost everything
James Stephen Brown (james-brown) · 2024-11-23T18:34:13.088Z · comments (0)
← previous page (newer posts) · next page (older posts) →