LessWrong 2.0 Reader

View: New · Old · Top

← previous page (newer posts) · next page (older posts) →

Loose thoughts on AGI risk
Yitz (yitz) · 2022-06-23T01:02:24.938Z · comments (3)
Air Conditioner Test Results & Discussion
johnswentworth · 2022-06-22T22:26:26.643Z · comments (42)
Announcing the LessWrong Curated Podcast
Ben Pace (Benito) · 2022-06-22T22:16:58.170Z · comments (27)
Google's new text-to-image model - Parti, a demonstration of scaling benefits
Kayden (kunvar-thaman) · 2022-06-22T20:00:59.930Z · comments (4)
Building an Epistemic Status Tracker
rcu · 2022-06-22T18:57:34.198Z · comments (6)
[link] Confusion about neuroscience/cognitive science as a danger for AI Alignment
Samuel Nellessen (samuel-nellessen) · 2022-06-22T17:59:31.140Z · comments (1)
[question] How do I use caffeine optimally?
randomstring · 2022-06-22T17:59:18.259Z · answers+comments (31)
Make learning a reality
Dalton Mabery (dalton-mabery) · 2022-06-22T15:58:05.959Z · comments (2)
Reflection Mechanisms as an Alignment target: A survey
Marius Hobbhahn (marius-hobbhahn) · 2022-06-22T15:05:55.703Z · comments (1)
House Phone
jefftk (jkaufman) · 2022-06-22T14:20:06.586Z · comments (2)
How to Visualize Bayesianism
David Udell · 2022-06-22T13:57:09.721Z · comments (2)
[question] Are there spaces for extremely short-form rationality content?
Aleksi Liimatainen (aleksi-liimatainen) · 2022-06-22T10:39:30.259Z · answers+comments (1)
Solstice Movie Review: Summer Wars
JohnBuridan · 2022-06-22T01:09:26.749Z · comments (6)
Security Mindset: Lessons from 20+ years of Software Security Failures Relevant to AGI Alignment
elspood · 2022-06-21T23:55:39.918Z · comments (42)
[link] A Quick List of Some Problems in AI Alignment As A Field
NicholasKross · 2022-06-21T23:23:31.719Z · comments (12)
[question] What is the difference between AI misalignment and bad programming?
puzzleGuzzle · 2022-06-21T21:52:57.362Z · answers+comments (2)
[link] What I mean by the phrase “getting intimate with reality”
Luise · 2022-06-21T19:42:56.578Z · comments (0)
[link] What I mean by the phrase "taking ideas seriously"
Luise · 2022-06-21T19:42:56.547Z · comments (2)
Hydrophobic Glasses Coating Review
jefftk (jkaufman) · 2022-06-21T18:00:05.426Z · comments (6)
[link] Progress links and tweets, 2022-06-20
jasoncrawford · 2022-06-21T17:12:44.361Z · comments (2)
[link] Debating Whether AI is Conscious Is A Distraction from Real Problems
sidhe_they · 2022-06-21T16:56:04.474Z · comments (10)
Mitigating the damage from unaligned ASI by cooperating with aliens that don't exist yet
MSRayne · 2022-06-21T16:12:01.753Z · comments (7)
The inordinately slow spread of good AGI conversations in ML
Rob Bensinger (RobbBB) · 2022-06-21T16:09:57.859Z · comments (62)
Getting from an unaligned AGI to an aligned AGI?
Tor Økland Barstad (tor-okland-barstad) · 2022-06-21T12:36:13.928Z · comments (7)
Common but neglected risk factors that may let you get Paxlovid
DirectedEvolution (AllAmericanBreakfast) · 2022-06-21T07:34:02.685Z · comments (8)
Dagger of Detect Evil
lsusr · 2022-06-21T06:23:01.264Z · comments (20)
[question] How easy/fast is it for a AGI to hack computers/a human brain?
Noosphere89 (sharmake-farah) · 2022-06-21T00:34:34.590Z · answers+comments (1)
[question] What is the most probable AI?
Zeruel017 · 2022-06-20T23:26:01.467Z · answers+comments (0)
Evaluating a Corsi-Rosenthal Filter Cube
jefftk (jkaufman) · 2022-06-20T19:40:01.980Z · comments (3)
Survey re AIS/LTism office in NYC
RyanCarey · 2022-06-20T19:21:33.642Z · comments (0)
Is This Thing Sentient, Y/N?
Thane Ruthenis · 2022-06-20T18:37:59.380Z · comments (9)
Steam
abramdemski · 2022-06-20T17:38:58.548Z · comments (13)
Parable: The Bomb that doesn't Explode
Lone Pine (conor-sullivan) · 2022-06-20T16:41:14.611Z · comments (5)
On corrigibility and its basin
Donald Hobson (donald-hobson) · 2022-06-20T16:33:06.286Z · comments (3)
Announcing the DWATV Discord
Zvi · 2022-06-20T15:50:03.051Z · comments (9)
Key Papers in Language Model Safety
aogara (Aidan O'Gara) · 2022-06-20T15:00:59.858Z · comments (1)
Relationship Advice Repository
Ruby · 2022-06-20T14:39:36.548Z · comments (36)
Adaptation Executors and the Telos Margin
Plinthist (Kredo) · 2022-06-20T13:06:29.519Z · comments (8)
Are we there yet?
theflowerpot · 2022-06-20T11:19:56.253Z · comments (2)
Causal confusion as an argument against the scaling hypothesis
RobertKirk · 2022-06-20T10:54:05.623Z · comments (30)
An AI defense-offense symmetry thesis
Chris van Merwijk (chrisvm) · 2022-06-20T10:01:18.968Z · comments (9)
Let's See You Write That Corrigibility Tag
Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2022-06-19T21:11:03.505Z · comments (69)
Half-baked alignment idea: training to generalize
Aaron Bergman (aaronb50) · 2022-06-19T20:16:43.735Z · comments (2)
Where I agree and disagree with Eliezer
paulfchristiano · 2022-06-19T19:15:55.698Z · comments (219)
[question] AI misalignment risk from GPT-like systems?
fiso64 (fiso) · 2022-06-19T17:35:41.095Z · answers+comments (8)
[Link-post] On Deference and Yudkowsky's AI Risk Estimates
bmg · 2022-06-19T17:25:14.537Z · comments (8)
Have The Effective Altruists And Rationalists Brainwashed Me?
UtilityMonster (Matt Goldwater) · 2022-06-19T16:05:04.380Z · comments (2)
Hebbian Learning Is More Common Than You Think
Aleksi Liimatainen (aleksi-liimatainen) · 2022-06-19T15:57:08.378Z · comments (2)
[link] The Malthusian Trap: An Extremely Short Introduction
Davis Kedrosky · 2022-06-19T15:25:44.026Z · comments (0)
Parliaments without the Parties
Yair Halberstadt (yair-halberstadt) · 2022-06-19T14:06:23.167Z · comments (18)
← previous page (newer posts) · next page (older posts) →