LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

Ball Square Station and Ridership Maximization
jefftk (jkaufman) · 2023-01-06T13:20:02.230Z · comments (0)
200 COP in MI: Techniques, Tooling and Automation
Neel Nanda (neel-nanda-1) · 2023-01-06T15:08:27.524Z · comments (0)
Definitions of “objective” should be Probable and Predictive
Rohin Shah (rohinmshah) · 2023-01-06T15:40:30.813Z · comments (27)
Categorizing failures as “outer” or “inner” misalignment is often confused
Rohin Shah (rohinmshah) · 2023-01-06T15:48:51.739Z · comments (21)
AI security might be helpful for AI alignment
Igor Ivanov (igor-ivanov) · 2023-01-06T20:16:40.446Z · comments (1)
[link] Why didn't we get the four-hour workday?
jasoncrawford · 2023-01-06T21:29:38.995Z · comments (34)
The Limit of Language Models
DragonGod · 2023-01-06T23:53:32.638Z · comments (26)
[link] [Linkpost] Jan Leike on three kinds of alignment taxes
Akash (akash-wasil) · 2023-01-06T23:57:34.788Z · comments (2)
Implications of simulators
ThomasW (ThomasWoodside) · 2023-01-07T00:37:45.799Z · comments (0)
[question] [Discussion] How Broad is the Human Cognitive Spectrum?
DragonGod · 2023-01-07T00:56:21.456Z · answers+comments (51)
Optimizing Human Collective Intelligence to Align AI
Shoshannah Tekofsky (DarkSym) · 2023-01-07T01:21:25.328Z · comments (5)
An Ignorant View on Ineffectiveness of AI Safety
Iknownothing · 2023-01-07T01:29:59.126Z · comments (7)
Benevolent AI and mental health
peter schwarz (peter-schwarz) · 2023-01-07T01:30:08.896Z · comments (2)
Can ChatGPT count?
p.b. · 2023-01-07T07:57:20.949Z · comments (11)
Anchoring focalism and the Identifiable victim effect: Bias in Evaluating AGI X-Risks
Remmelt (remmelt-ellen) · 2023-01-07T09:59:52.120Z · comments (2)
Open & Welcome Thread - January 2023
DragonGod · 2023-01-07T11:16:18.646Z · comments (37)
What should a telic science look like?
mrcbarbier · 2023-01-07T16:13:28.849Z · comments (0)
Building a Rosetta stone for reductionism and telism (WIP)
mrcbarbier · 2023-01-07T16:22:55.465Z · comments (0)
What's wrong with the paperclips scenario?
No77e (no77e-noi) · 2023-01-07T17:58:35.866Z · comments (11)
The Fountain of Health: a First Principles Guide to Rejuvenation
PhilJackson · 2023-01-07T18:34:12.094Z · comments (38)
[question] Asking for a name for a symptom of rationalization
metachirality · 2023-01-07T18:34:30.364Z · answers+comments (5)
[link] Nothing New: Productive Reframing
adamShimi · 2023-01-07T18:43:35.617Z · comments (7)
Looking for Spanish AI Alignment Researchers
Antb (darustc4) · 2023-01-07T18:52:08.375Z · comments (3)
[link] David Krueger on AI Alignment in Academia, Coordination and Testing Intuitions
Michaël Trazzi (mtrazzi) · 2023-01-07T19:59:09.785Z · comments (0)
Protectionism will Slow the Deployment of AI
bgold · 2023-01-07T20:57:11.644Z · comments (6)
[question] How do I better stick to a morning schedule?
Randomized, Controlled (BossSleepy) · 2023-01-08T00:52:52.791Z · answers+comments (8)
Feature suggestion: add a 'clarity score' to posts
LVSN · 2023-01-08T01:00:29.029Z · comments (5)
Can Ads be GDPR Compliant?
jefftk (jkaufman) · 2023-01-08T02:50:01.486Z · comments (10)
[link] Stop Talking to Each Other and Start Buying Things: Three Decades of Survival in the Desert of Social Media
the gears to ascension (lahwran) · 2023-01-08T04:45:11.413Z · comments (14)
AI psychology should ground the theories of AI consciousness and inform human-AI ethical interaction design
Roman Leventov · 2023-01-08T06:37:54.090Z · comments (8)
Could evolution produce something truly aligned with its own optimization standards? What would an answer to this mean for AI alignment?
No77e (no77e-noi) · 2023-01-08T11:04:40.642Z · comments (4)
Dangers of deference
TsviBT · 2023-01-08T14:36:33.454Z · comments (5)
Halifax Monthly Meetup: Moloch in the HRM
Ideopunk · 2023-01-08T14:49:28.054Z · comments (0)
200 COP in MI: Image Model Interpretability
Neel Nanda (neel-nanda-1) · 2023-01-08T14:53:14.681Z · comments (3)
[question] Research ideas (AI Interpretability & Neurosciences) for a 2-months project
flux (FoxYou) · 2023-01-08T15:36:12.984Z · answers+comments (1)
[question] What specific thing would you do with AI Alignment Research Assistant GPT?
quetzal_rainbow · 2023-01-08T19:24:26.221Z · answers+comments (9)
I tried to learn as much Deep Learning math as I could in 24 hours
Phosphorous (everett-smith) · 2023-01-08T21:07:34.190Z · comments (2)
Citability of Lesswrong and the Alignment Forum
Leon Lang (leon-lang) · 2023-01-08T22:12:02.046Z · comments (2)
Latent variable prediction markets mockup + designer request
tailcalled · 2023-01-08T22:18:36.050Z · comments (4)
[question] GPT learning from smarter texts?
Viliam · 2023-01-08T22:23:26.131Z · answers+comments (7)
Simulacra are Things
janus · 2023-01-08T23:03:26.052Z · comments (7)
ChatGPT tells stories about XP-708-DQ, Eliezer, dragons, dark sorceresses, and unaligned robots becoming aligned
Bill Benzon (bill-benzon) · 2023-01-08T23:21:18.748Z · comments (2)
On Blogging and Podcasting
DanielFilan · 2023-01-09T00:40:00.908Z · comments (6)
You're Not One "You" - How Decision Theories Are Talking Past Each Other
keith_wynroe · 2023-01-09T01:21:11.708Z · comments (11)
Wolf Incident Postmortem
jefftk (jkaufman) · 2023-01-09T03:20:03.723Z · comments (13)
Rationality Practice: Self-Deception
Darmani · 2023-01-09T04:07:03.845Z · comments (0)
Wearable tech might disrupt language before vision
Sheikh Abdur Raheem Ali (sheikh-abdur-raheem-ali) · 2023-01-09T05:59:18.089Z · comments (0)
[link] Big list of AI safety videos
JakubK (jskatt) · 2023-01-09T06:12:35.139Z · comments (2)
Reification bias
adamShimi · 2023-01-09T12:22:15.460Z · comments (6)
← previous page (newer posts) · next page (older posts) →