Longtermism vs short-termism for personal life extension 2021-07-17T03:52:54.407Z
If instead of giving out dividends, public companies bought total market index funds, which companies would be the biggest? 2021-06-15T17:07:05.338Z
What's your probability that the concept of probability makes sense? 2021-05-22T21:53:31.766Z
[May – weekly] Focusmate: identity preservation 2021-04-25T17:58:28.897Z
Is there any plausible mechanisms for why taking an mRNA vaccine might be undesirable for a young healthy adult? 2021-04-08T01:18:42.434Z
What do you think would be the best investment policy for a cryonics trust? 2021-04-04T19:20:22.705Z
2 innovative life extension approaches using cryonics technology 2021-04-02T00:54:04.539Z
[April – weekly] Focusmate: biostasis, lifelogging, death-focus ethics 2021-03-31T16:49:02.604Z
[event] Cryonics icebreaker 2021-03-21T17:44:29.114Z
What are fun little puzzles / games / exercises to learn interesting concepts? 2021-03-18T03:26:00.948Z
[recurrent] Focusmate: cryonics and lifelogging 2021-03-11T16:28:34.263Z
Welcome to LW/SSC/EA in San Juan, Puerto Rico 2021-03-09T18:49:35.733Z
What ethical thought experiments can be reversed? 2021-03-06T15:12:09.826Z
Focusmate: Cryonics sign up 2021-03-01T15:41:10.366Z
How much harder is it to revive a neuro-only cryonics patient? 2021-01-12T23:24:45.963Z
Signaling importance 2020-12-08T09:14:36.148Z
Predictions made by Mati Roy in early 2020 2020-11-21T03:24:56.020Z
What fraction of Dan Ariely's Irrational Game hasn't replicated? 2020-11-09T20:25:27.445Z
What features would you like a prediction platform to have? 2020-10-13T00:48:03.024Z
Reviews of the book 'The Alignment Problem' 2020-10-11T07:41:14.841Z
Reviews of TV show NeXt (about AI safety) 2020-10-11T04:31:48.363Z
Buying micro-biostasis 2020-10-07T10:55:19.006Z
What reacts would you like to be able to give on posts? (emoticons, cognicons, and more) 2020-10-04T18:31:06.596Z
What are examples of Rationalist fable-like stories? 2020-09-28T16:52:13.500Z
What are good ice breaker questions for meeting people in this community? 2020-09-28T15:07:16.798Z
What hard science fiction stories also got the social sciences right? 2020-09-27T20:37:44.256Z
Surviving Petrov Day 2020-09-26T16:40:03.169Z
Has anyone written stories happening in Hanson's em world? 2020-09-21T14:37:11.150Z
For what X would you be indifferent between living X days, but forgetting your day at the end of everyday, or living 10 days? (terminally) 2020-09-18T04:05:59.078Z
How do you celebrate your birthday? 2020-09-17T10:00:50.609Z
What are examples of simpler universes that have been described in order to explain a concept from our more complex universe? 2020-09-17T01:31:10.367Z
What are examples of 'scientific' studies that contradict what you believe about yourself? 2020-08-03T06:11:19.683Z
When a status symbol loses its plausible deniability, how much power does it lose? 2020-07-07T00:48:21.558Z
The Echo Fallacy 2020-07-05T23:00:39.476Z
[Crowdfunding] LessWrong podcast 2020-07-03T20:59:53.590Z
The Book of HPMOR Fanfics 2020-07-03T13:32:17.536Z
Is taking bacopa good for life extension? 2020-05-23T08:54:27.480Z
What aspects of the world emotionally bothers you on an immediate personal level on a daily basis? 2020-05-22T06:27:55.357Z
[link] Biostasis / Cryopreservation Survey 2020 2020-05-16T07:20:58.879Z
What was your reasoning for deciding whether to raise children? 2020-05-15T03:53:23.776Z
What work of fiction explore increased transparency in the world? 2020-05-13T21:15:01.640Z
What are articles on "lifelogging as life extension"? 2020-05-13T20:35:10.676Z
Is AI safety research less parallelizable than AI research? 2020-05-10T20:43:59.476Z
What are examples of perennial discoveries? 2020-05-09T06:17:18.171Z
What fraction of your lifetime (0-80 years old) egoist budget would you (want to) spend on a pill that made you live for as long as you wanted (perfect invincibility), as healthily as you wanted if you knew it would become available to you once you're 80 years old (and that you would otherwise irreversibly die)? 2020-05-08T08:50:40.815Z
What would you do differently if you were less concerned with looking weird? 2020-05-07T23:29:16.618Z
Why do you (not) use a pseudonym on LessWrong? 2020-05-07T19:34:35.446Z
[link] How many humans will have their brain preserved? Forecasts and trends 2020-05-07T06:11:00.176Z
How much money would you pay to get access to video footage of your surroundings for a year of your choice (in the past)? 2020-05-05T05:50:44.628Z
[Link] Rationalist Games -- Facebook Group 2020-05-04T22:13:41.748Z


Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-23T21:22:15.482Z · LW · GW

Hobby: serve so many bullets to sophisticated philosophers that they're missing half their teeth by the end of the discussion

Comment by Mati_Roy (MathieuRoy) on Longtermism vs short-termism for personal life extension · 2021-07-17T16:35:18.155Z · LW · GW

I love this, thank you!

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-15T17:12:21.131Z · LW · GW

afaik, most board members are very passive, and hasnt been doing the things Nicolas wants to do

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-15T17:07:47.397Z · LW · GW
Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-14T03:33:58.259Z · LW · GW

For the Cryonics Institute board elections, I recommend voting for Nicolas Lacombe.

I’ve been friends with Nicolas for over a decade. Ze’s very principled, well organized and hard working. I have high trust in zir, and high confidence ze would be a great addition to the CI's board.

I recommend you cast some or all of your votes for Nicolas (you can cast up to 4 votes total). If you’re signed up with CI, simply email with your votes.

see zir description here:

Comment by Mati_Roy (MathieuRoy) on The case for lifelogging as life extension · 2021-07-11T16:20:35.838Z · LW · GW

my updated guide:

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-07T20:32:04.046Z · LW · GW

before the birth of that AI, we could split the Universe among existing beings

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-07-02T03:57:39.154Z · LW · GW

crazy idea i just had mayyybe a deontological Libertarian* AI with (otherwise) any utility function is not that bad (?) maybe that should be one of the thing we try (??????) *where negative externality also count as aggressions, and other such fixes to naive libertarianism

Comment by Mati_Roy (MathieuRoy) on How much interest would there be in a fringe theories wiki? · 2021-06-29T17:40:20.233Z · LW · GW

I probably would mostly still use LessWrong for the wiki for my writings.

But as a reader, I would probably read at least the summary of each article, and probably the full text of each potentially actionable-for-me articles.

But I'd say finding the writers is harder than finding the readers

Another idea would be to have a "fringe" tag on LessWrong. But, actually, I think I would choose a more connotationally neutral term, like "speculative". Although, still, that's a changing and subjective category, so not sure I would do that.

Another idea would be to just curate fringe articles (like you've done in the description of this question). Maybe ask a question "What are articles on interesting fringe ideas?" on LessWrong.

You could also have a yearly prize of like 100$ to submit articles on fringe ideas. This might be a more cost-effective way to achieve your goal.

Comment by Mati_Roy (MathieuRoy) on Cryonics signup guide #1: Overview · 2021-06-25T03:55:45.546Z · LW · GW

ah, I see! yeah, I guess being an associate member is stronger evidence (but am almost certain some people were preserved without this)

Comment by Mati_Roy (MathieuRoy) on Let's create a market for cryonics · 2021-06-15T16:13:12.168Z · LW · GW

I wonder what's the life expectancy of the average cryonicist taking a life insurance vs the general population taking a life insurance. If it's higher, then cryonics-purposed life insurances could be cheaper. The first insurance company to offer this would grab a big part of the cryonics insurance market. Life insurance might for once live up to its name :)

Comment by Mati_Roy (MathieuRoy) on Forecasting Thread: AI Timelines · 2021-06-14T21:54:39.411Z · LW · GW

Without consulting my old prediction here, I answered someone asking me:

What is your probability mass for the date with > 50% chance of agi?


I used to use the AGI definition "better and cheaper than humans at all economic tasks", but now I think even if we're dumber, we might still be better at some economic tasks simply because we know human values more. Maybe the definition could be "better and cheaper at any well defined tasks". In that case, I'd say maybe 2080, taking into account some probability of economic stagnation and some probability that sub-AGI AIs cause an existential catastrophe (and so we don't develop AGI)

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-06-14T17:06:26.464Z · LW · GW

In the book Superintelligence, box 8, Nick Bostrom says:

How an AI would be affected by the simulation hypothesis depends on its values. [...] consider an AI that has a more modest final goal, one that could be satisfied with a small amount of resources, such as the goal of receiving some pre-produced cryptographic reward tokens, or the goal of causing the existence of forty-five virtual paperclips. Such an AI should not discount those possible worlds in which it inhabits a simulation. A substantial portion of the AI’s total expected utility might derive from those possible worlds. The decision-making of an AI with goals that are easily resource-satiable may therefore—if it assigns a high probability to the simulation hypothesis—be dominated by considerations about which actions would produce the best result if its perceived world is a simulation. Such an AI (even if it is, in fact, not in a simulation) might therefore be heavily influenced by its beliefs about which behaviors would be rewarded in a simulation. In particular, if an AI with resource-satiable final goals believes that in most simulated worlds that match its observations it will be rewarded if it cooperates (but not if it attempts to escape its box or contravene the interests of its creator) then it may choose to cooperate. We could therefore find that even an AI with a decisive strategic advantage, one that could in fact realize its final goals to a greater extent by taking over the world than by refraining from doing so, would nevertheless balk at doing so.

  1. If the easily resource-satiable goals are persistent through time (ie. the AI wants to fulfill them for the longest period of time possible), then the AI will either try to keep the simulation running for as long as possible (and so not grab its universe) or try to escape the simulation.

  2. If the easily resource-satiable goals are NOT persistent through time (ie. once the AI has created the 45 virtual paperclips, it doesn't matter if they get deleted, the goal has already been achieved), then once the AI has created the 45 paperclips, it has nothing to lose by grabbing more resources (gradually, until it has grabbed the Universe), but it has something to win, namely: a) increasing its probability (arbitrarily close to 100%) that it did in fact achieve its goal through further experiment and reasoning (ie. because it could be mistaken about having created 45 virtual paperclips), and b) if it didn't, then remedy to that.

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-06-14T16:54:09.750Z · LW · GW


Comment by Mati_Roy (MathieuRoy) on Cryonics signup guide #1: Overview · 2021-06-12T22:27:34.964Z · LW · GW

Alcor offers worldwide standby services

Comment by Mati_Roy (MathieuRoy) on Cryonics signup guide #1: Overview · 2021-06-12T22:26:10.148Z · LW · GW

Just in case anyone cares: There are ways you can increase your own chances of a good preservation, notably by moving near Alcor.

Comment by Mati_Roy (MathieuRoy) on Cryonics signup guide #1: Overview · 2021-06-12T22:19:21.608Z · LW · GW

I was gonna point out the same thing

Comment by Mati_Roy (MathieuRoy) on Cryonics signup guide #1: Overview · 2021-06-12T22:14:44.220Z · LW · GW

I'm pretty sure paying a monthly fee is not required to have informed consent. Can you quote the part of the text that says otherwise?

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-06-12T22:12:59.221Z · LW · GW

thanks! yeah i know, but would like if it was more easily accessible whenever i watch a video:)

Comment by Mati_Roy (MathieuRoy) on #2: Neurocryopreservation vs whole-body preservation · 2021-06-02T13:21:59.076Z · LW · GW

Seen on the Facebook group:

Dora Kent still has a chance at resurrection because she was a neuro patient. Had she been whole body, the Riverside coroner would have sliced her brain to pieces. Fortunately, before the Coroner executed a search warrant, her head mysteriously disappeared from the Alcor facility. That gave Alcor the time to get a permanent injunction in the courts against autopsying her head.

They seek it here... They seek it there... Those coroners seek it everywhere. Is it alive or is it dead? That damn elusive frozen head.

Frozen heads are a whole lot easier to move in the event of an emergency, be it legal, criminal, war, natural disasters or whatever. Costs a lot less to keep them cool as well. Looking at the long haul, and given that cryonics is a highly speculative endeavor that will likely require almost unimaginable technology to work, it's a rational choice.

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-27T20:36:28.407Z · LW · GW


Comment by Mati_Roy (MathieuRoy) on Sabien on "work-life" balance · 2021-05-27T15:42:31.763Z · LW · GW

investing-consumption ratio

related: "Consume now or later?":

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-27T15:37:07.543Z · LW · GW

oh, LW actually has a bookmark feature, which i could use for that! although i prefer using it for articles i want to read

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-27T05:34:04.716Z · LW · GW

can you say more about agriculture?

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-25T16:52:47.868Z · LW · GW

i want to invest in companies that will increase in value if AI capabilities increases fast / faster than what the market predicts

do you have suggestions?

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-24T03:33:56.014Z · LW · GW

a feature i would like on content website like YouTube and LessWrong is an option to market a video/article as read as a note to self (x-post fb)

Comment by Mati_Roy (MathieuRoy) on On the unpopularity of cryonics: life sucks, but at least then you die · 2021-05-21T03:24:54.924Z · LW · GW

"life's a bitch and then you die" -Young Sinatra III (

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-05-08T19:03:20.044Z · LW · GW

If crypto makes the USD go to 0, will life insurances denominated in USD not have anything to pay out? Maybe an extra reason for cryonicists to own some crypto


Comment by Mati_Roy (MathieuRoy) on Is there any plausible mechanisms for why taking an mRNA vaccine might be undesirable for a young healthy adult? · 2021-04-15T21:38:09.518Z · LW · GW

Epistemic status: someone told me that someone that knew about this told them that

mRNA vaccines have been done on non-human animals for decades, and they seem fine

Comment by Mati_Roy (MathieuRoy) on Is there any plausible mechanisms for why taking an mRNA vaccine might be undesirable for a young healthy adult? · 2021-04-10T04:30:20.877Z · LW · GW

well, I should unpack "fully isolate" (I meant fully isolate from the virus):

  • I isolate with another human
  • I video-chat with multiple humans
  • I chat in person with other humans outside with 3 meters distance or insider with a full-face mask :)
Comment by Mati_Roy (MathieuRoy) on What do you think would be the best investment policy for a cryonics trust? · 2021-04-04T19:30:29.262Z · LW · GW

Am just thinking about this now, so not high confidence.

I'm thinking a total market cap index of all publicly traded companies in the world -- unless some stock exchanges are really sketchy for some reasons (I'm aware "sketchy" is ill-defined).

Also maybe a fraction (10%?) stored in a total market cap of deflationary currencies (decentralized ones I suppose, as for centralized ones it's generally harder to tell if they will become inflationary) as they will presumably become more in demand, in the short term as the economy might transition to them, but also in the long term as the population grows and so those the demand for currency (although if minds start running faster, I think that would decrease the value of currencies as they would be traded more often).

Maybe worlds in which reanimation is possible is correlated with other technological development, and so that might also inform some investments, but not sure which ones, and that starts adding a lot of complexity, at least while we don't have large prediction markets.

I'm also thinking those investment don't need to be liquid at all, and REITs often have a higher return because they are less liquid than stocks (AFAIU); so if this is true, then maybe having a lot of investments in REITs would also be beneficial. (2021-05-08 update: this is also true of other investment vehicles, like loans)

Comment by Mati_Roy (MathieuRoy) on [event] Cryonics icebreaker · 2021-04-04T17:11:37.499Z · LW · GW

Those are the questions we intend to use for the icebreakers. No need to read them in advance -- just leaving them here in case anyone is interested.

  • What do you think lead you to be interested in or curious about cryonics?
  • What's your probability that cryonics works? Why?
  • When was the last time you had a significant change in your worldview?
  • What makes you excited?
  • What's your origin story? (or how did you discover cryonics)
  • What's a belief you have that you think your cryonics peers would find weird?
  • What's your favorite Trolley problem variant?
  • What makes you excited?
  • What's an interesting reaction you got from talking about cryonics?
  • What's your lying policy? (ie. when do you find it acceptable to lie)
  • What do you think are plausible revival scenarios?
  • What makes you sad?
  • What was your family’s reaction or how do you think they would react?
  • What's a belief you have that you think most people would find weird?
  • What do you think of as the biggest obstacle to revival? A. Technology required for repair B. Long term stability of the organizations C. Chance of suboptimal preservation (information loss) D. Other
  • What makes you anxious?
  • If we're living in a simulation, would you like to get out of it right now or wait longer?
  • What are your revival timelines?
  • The zombie apocalypse is coming, who are 3 people you want on your team?
  • What gives you wonder?

For the group activity:

  • What do you think is the best way to market cryonics?

Possible alternatives:

----> What is the best way to increase cryonics membership?

----> What is the best way to popularize cryonics?

Comment by Mati_Roy (MathieuRoy) on 2 innovative life extension approaches using cryonics technology · 2021-04-02T15:52:44.919Z · LW · GW

I like my comment on the Effective Altruism & Life Extension post ^ _ ^

the reason I posted this article here yesterday is because of the problem that once cryonics will be adopted globally, people will want to get pre-emptively cryopreserved to avoid any further identity degradation. however, if everyone does that, it becomes an existential risk (because no one will be around to keep improving our tech). it's a massive ultimate group prisoner dilemma. it's in anyone's best interest to do it themselves, but we would rather no one (else) does it. and it's a hard coordination problem. split brain preservation solves this by allowing the whole population to preserve half of their brain while using their other half to keep working towards the goal of reanimating the preserved half. oh, also, happy April Fools day!

Comment by Mati_Roy (MathieuRoy) on 2 innovative life extension approaches using cryonics technology · 2021-04-02T15:31:58.862Z · LW · GW

Aschwin de Wolf commented on my post in the cryonics Facebook group with interesting information before knowing it was posted in the spirit of April's Fool ^ _ ^ :

Idea 1 was already proposed by Robert Prehoda in his book Suspended Animation in 1969. Practicalities aside, this is guaranteed to lead to significant cognitive damage after several sessions because it will only require one botched hypothermic circulatory arrest procedure to render the person a vegetable. In addition, the kind of broad recovery shown in animals falls short of the finer cognitive recovery that is a condition for this approach to even make sense at a technical level. Just dive a little deeper into the literature about profound hypothermic and ultra-profound hypothermic circulatory arrest.

Idea 2 is probably meant seriously but will just invite ridicule by mainstream observers and scientists by combining extreme reductionism about the nature of identity and looking really, really, desperate. You may gain two "rationalist" advocates but alienate a lot more.

Idea 2 also start off with: "A big dilemma in the biostasis community is whether one should get cryopreserved now or later." I am not aware of any of my colleagues considering this a "big dilemma" unless in cases of an early dementia diagnosis.

Comment by Mati_Roy (MathieuRoy) on 2 innovative life extension approaches using cryonics technology · 2021-04-02T01:53:35.092Z · LW · GW

woops, I just realized the link was missing:

Comment by Mati_Roy (MathieuRoy) on Selling Attention for Money · 2021-03-27T00:48:23.147Z · LW · GW

Offer: n_th card you personally buy is n USD (doesn't matter if other people bought cards before you)

There's a capped to this offer, but it's probably high enough that it won't be reached

I will check the deck at least once per week for at least 16 weeks (if I miss a week, the timespan is increased by a week; if still not done a year from now, then money back; but that shouldn't happen).

I invite you to abuse this, but I also reserve the right to refuse transactions:)

I might accept anonymous buyers; if you're interested, talk to me here:

payment can be made to:

  • PayPal
  • Bitcoin wallet: 3C7J2X3LswsmGewPi4a63jmiFAoUzCu3vE
  • Eth wallet: 0x4d06c5CdeaF69f76677629296a8C3D3B6901638D
  • other payment methods available on request

Additional notes:

Fermi estimate of time requirement: checking each cards 15 times for 15 seconds = 3m45s, so roughly 3-10 USD per card. The influence on my mind for cards I'll accept will probably mostly 'cost' me -10 to +15 USD. I also value the information I will gain about how this system plays out.

Comment by Mati_Roy (MathieuRoy) on MikkW's Shortform · 2021-03-24T02:47:02.296Z · LW · GW

That's genius! Can I (or you) create a LessWrong thread inviting others to do the same?

Comment by Mati_Roy (MathieuRoy) on MikkW's Shortform · 2021-03-24T02:44:24.296Z · LW · GW

if so, I want to claim 7 cards

Comment by Mati_Roy (MathieuRoy) on MikkW's Shortform · 2021-03-24T02:41:15.076Z · LW · GW

can I claim cards before choosing its content?

Comment by Mati_Roy (MathieuRoy) on Demand offsetting · 2021-03-22T17:22:27.212Z · LW · GW

documented here:

Comment by Mati_Roy (MathieuRoy) on [event] Cryonics icebreaker · 2021-03-21T17:54:59.155Z · LW · GW

Eric Magro had the genius idea of naming the event "M22 Icebreaker" given M22 is Alcor's cryoprotectant, which prevents ice formation ^^ We decided to go with a clearer title, but still wanted to share here 😅

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-03-21T05:14:39.792Z · LW · GW

ah, yeah, you're right! thank you

Comment by Mati_Roy (MathieuRoy) on Conspicuous saving · 2021-03-21T00:27:30.214Z · LW · GW

same is true with time; hobbies might be a lot about signaling wealth ("I'm so rich I can do non-productive activities"); retiring early actually sounds like a good signaling mechanism for this; retiring early likely(??) have better externalities than conspicuous consumptions, but ideally you would want people to be able to fully signal their wealth at 0 cost (whether time or monetary), and so retiring early for signaling purposes still seems to have a negative externality

another way to signal wealth might be by having a lot of children; and that one seems like it has positive externalities, and so removing the need for signaling wealth might reduce number of children

I'd like to see studies on how having your wealth publicly known changes those factors

Comment by Mati_Roy (MathieuRoy) on Conspicuous saving · 2021-03-21T00:19:55.398Z · LW · GW

Maybe we have strong instincts for conspicuous consumption for signaling purposes, so that even if on paper another mechanism should provide the same signaling value, maybe it wouldn't reduce conspicuous consumption as much as we'd expect (although possibly still worth it!)

Comment by Mati_Roy (MathieuRoy) on Conspicuous saving · 2021-03-21T00:17:12.283Z · LW · GW

Anyone has an idea how a company could provide such value? (ie. signaling wealth, with plausible deniability about the intentions, while investing the wealth)

Comment by Mati_Roy (MathieuRoy) on Mati_Roy's Shortform · 2021-03-21T00:15:24.064Z · LW · GW

the usual story is that Governments provide public good because Markets can't, but maybe Markets can't because Governments have secured a monopoly on them?


Comment by Mati_Roy (MathieuRoy) on Conspicuous saving · 2021-03-20T23:59:35.117Z · LW · GW

This hypothesis predicts that people who have a publicly known wealth spend less, controlling for other factors. Maybe we should start by studying this. (Seems high expected value to me!)

Comment by Mati_Roy (MathieuRoy) on Focusmate: Cryonics sign up · 2021-03-09T18:26:25.122Z · LW · GW

it did; we were 6; I will make it weekly

Comment by Mati_Roy (MathieuRoy) on Focusmate: Cryonics sign up · 2021-03-09T18:25:59.444Z · LW · GW

humm, should have been able to join during the whole hour

Comment by Mati_Roy (MathieuRoy) on What ethical thought experiments can be reversed? · 2021-03-07T14:00:17.717Z · LW · GW

Ahhh, yep, thanks