Posts

EAG After-After-Party 2024-11-01T00:13:10.818Z
Political Prereqs 2024-10-08T01:09:15.574Z
Feedback and. . . oops, it's time to overthrow the organizers 2024-08-05T23:31:20.480Z
Making Beliefs Pay Rent 2024-07-28T17:59:52.101Z
Is there a Schelling point for group house room listings? 2024-07-23T03:03:29.639Z
Room Available in Boston Group House 2024-07-23T02:55:59.602Z
Steelmanning the Devil 2024-03-05T05:14:15.279Z
Humans Can Be Manually Strategic 2024-01-06T19:28:44.571Z
Units of Exchange 2023-09-03T18:46:56.021Z
Aversion Factoring Meetup 2023-07-06T04:15:16.181Z
Stanford claims to have replicated ChatGPT for < $600 2023-03-21T02:28:43.576Z
ACX Prediction Contest Discussion 2023-01-05T02:53:27.379Z
Is there a clearly laid-out write-up of the case to drop Covid precautions? 2021-11-09T02:46:41.387Z
A Rational New Year 2020-12-21T05:51:53.432Z
Cambridge, MA Rationalist Reading Group 2020-10-19T04:01:24.060Z
How much to worry about the US election unrest? 2020-10-12T03:39:55.105Z
Reading Discussion Group 2020-09-29T03:59:14.510Z
C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-09-29T03:42:07.369Z
Reading Discussion Group 2020-09-06T18:49:15.044Z
C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-08-28T02:45:11.504Z
Rationalist Reading Group (Online) 2020-07-30T02:14:04.411Z
C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-07-23T03:25:25.510Z
C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-07-02T03:45:06.577Z
C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-06-02T03:17:49.849Z
̶C̶a̶m̶b̶r̶i̶d̶g̶e̶ Virtual LW/SSC Meetup 2020-04-27T03:41:39.390Z
Cambridge LW/SSC Meetup 2020-02-22T19:56:28.682Z
Cambridge LW/SSC Meetup: Prediction Training 2020-01-28T05:25:07.466Z
An Empistemically Rational Superbowl 2020-01-28T04:05:40.095Z
Cambridge Prediction Game 2020-01-25T03:57:59.721Z
Cambridge LW/SC Meetup 2019-12-28T03:07:07.008Z
Cambridge LW/SSC Meetup 2019-11-18T02:51:29.797Z
Cambridge LW/SSC Meetup 2019-10-30T13:35:14.247Z
Cambridge LW/SSC Meetup 2019-10-02T03:37:53.551Z
SSC Meetups Everywhere 2019-09-10T06:02:10.679Z
Cambridge LW/SSC Meetup 2019-08-31T17:10:38.696Z
Cambridge LW/SSC Meetup 2019-07-23T00:47:42.743Z
Self-experiment Protocol: Effect of Chocolate on Sleep 2019-07-22T03:32:19.106Z
Cambridge LW/SSC Meetup 2019-06-03T01:55:57.895Z
Cambridge LessWrong / SSC Meetup 2019-05-13T01:56:43.913Z
Cambridge SSC Meetup 2019-04-05T02:12:00.121Z
Cambridge SSC Meetup 2019-02-17T18:28:46.057Z
Cambridge SSC Meetup 2019-02-17T18:27:07.227Z
Cambridge SlateStarCodex Meetup 2019-01-06T05:11:20.893Z
Why Most Intentional Communities Fail (And Some Succeed) 2017-05-22T03:04:23.288Z
What are you surprised people pay for instead of doing themselves? 2017-02-13T01:07:33.803Z
What are you surprised people don't just buy? 2017-02-13T01:07:12.335Z
[Link] White House announces a series of workshops on AI, expresses interest in safety 2016-05-04T02:50:22.434Z
Group Rationality Diary, February 2016 2016-02-14T01:55:15.553Z
Study partner matching thread 2016-01-25T04:25:03.592Z
[Link] Stephen Hawking AMA answers 2015-10-08T23:13:34.610Z

Comments

Comment by NoSignalNoNoise (AspiringRationalist) on Heresies in the Shadow of the Sequences · 2024-11-16T03:48:28.095Z · LW · GW

I like what you're doing, but I feel like the heresies you propose are too tame.

Here are some more radical heresies to consider:

  1. Most people are far more bottlenecked on some combination of akrasia and prospective memory, not on the accuracy of their models of the world. Rationalists in particular would be better off devoting effort to actually doing the obvious things than to understanding the world better.
  2. Self deception is very instrumentally useful a large fraction of real world situations we find ourselves in, and we should use more of it.
    1. Mormons seem to be especially good at coordinating on good lifestyle choices, so we should all consider becoming Mormon.
  3. Among groups of 10+ people, it's usually more useful to get everyone all working on implementing the same plan than it is to come up with the best plan.
  4. Intelligence (of the sort measured by exams and IQ tests) is only moderately important to success.
Comment by NoSignalNoNoise (AspiringRationalist) on Physical Therapy Sucks (but have you tried hiding it in some peanut butter?) · 2024-09-11T14:12:53.888Z · LW · GW

I generally watch videos I enjoy while doing physical therapy exercises. I didn't conceptualize it as hiding the "reward" from myself as an incentive for exercising; I conceptualize it as making the rather boring, sometimes aversive activity less salient by focusing my attention on something else.

As an example, I find it much easier to hold a plank when I'm focused on the video I'm watching than when I'm just starting at the timer counting down.

Comment by NoSignalNoNoise (AspiringRationalist) on Physical Therapy Sucks (but have you tried hiding it in some peanut butter?) · 2024-09-11T14:01:58.867Z · LW · GW

I've tried this approach, and although it works well during the early part of the game, in the late game, a single turn can take 5-10 minutes, which is much less helpful as an exercise interlude.

Comment by NoSignalNoNoise (AspiringRationalist) on Physical Therapy Sucks (but have you tried hiding it in some peanut butter?) · 2024-09-11T14:00:21.647Z · LW · GW

I've found that watching videos I enjoy while doing PT exercises helps. A key component of this strategy was to get a laptop stand with an adjustable angle so that I can position my screen somewhere I can see it (different places depending on how I'm physically positioned for each exercise).

Comment by NoSignalNoNoise (AspiringRationalist) on Music in the AI World · 2024-08-18T20:29:13.924Z · LW · GW

I think the current diversity of music is largely caused by artists' different lived experiences. You feel something, this is important for you, you try to express that via music. As long as AIs don't have anything like "unique experiences" on the scale of humans, I'm not sure if they'll be able to create music that is that diverse (and thus interesting).

If the AI customized it for each listener (and does a good job), then music will reflect the unique experiences of the listeners, which would result in a more diverse range of music than music that only reflects the unique experiences of musicians.

Of course, we could end up in an awkward middle ground where AI only generates variations on a successful pop music formula, and it all becomes a bland mush. But I think in that case, people would just go back to human-generated music on Spotify and YouTube.

Comment by NoSignalNoNoise (AspiringRationalist) on New blog: Expedition to the Far Lands · 2024-08-17T19:54:53.532Z · LW · GW

I like this collection of concepts, but I feel like I may not be understanding them very well without examples. Do you have any cached examples?

Comment by NoSignalNoNoise (AspiringRationalist) on Rabin's Paradox · 2024-08-17T19:43:28.919Z · LW · GW

Sure, but that does imply that your marginal utility of money decreases that fast outside that domain.

Comment by NoSignalNoNoise (AspiringRationalist) on Rabin's Paradox · 2024-08-16T00:32:20.439Z · LW · GW

The assumption that the marginal utility of wealth decreases exponentially doesn't seem justified to me. Why not some other positive-but-decreasing function, such as 1/W (which yield a logarithmic utility function)?

What properties does the utility function need to have for this result to generalize, and are those priorities reasonable to assume?

Comment by NoSignalNoNoise (AspiringRationalist) on Dragon Agnosticism · 2024-08-02T23:04:36.274Z · LW · GW

To pick an uncontroversial example, imagine someone glomerizing in whether the Earth was flat or (approximately) spherical. That would signal that you're the sort of person who considered a spherical Earth to be a plausible hypothesis, which is almost as bad as actually believing it. All reasonable, right-thinking people, on the other hand, know that it's obviously flat and wouldn't even consider such nonsense.

Comment by NoSignalNoNoise (AspiringRationalist) on If You Can Climb Up, You Can Climb Down · 2024-07-31T01:35:54.381Z · LW · GW

How do you determine where it's ok for her to go barefoot?

Comment by NoSignalNoNoise (AspiringRationalist) on Housing Roundup #9: Restricting Supply · 2024-07-18T01:45:51.602Z · LW · GW

I think it's intentional. He's saying that reducing the fee from $23k to $230 would be an improvement.

Comment by NoSignalNoNoise (AspiringRationalist) on patent process problems · 2024-07-14T21:43:13.385Z · LW · GW

The FDA used to have a long backlog of drug approval applications due to understaffing. This problem was eventually addressed by the Prescription Drug User Fee Act, which established filling fees for drug approvals and a deadline to review new drug applications. The filling fees were a way to make the cost of additional staffing politically palatable, and are much less than the costs of approval delays.

Comment by NoSignalNoNoise (AspiringRationalist) on Ice: The Penultimate Frontier · 2024-07-14T02:24:36.300Z · LW · GW

How does this compare to the costs of making (part of) Antarctica habitable?

Comment by NoSignalNoNoise (AspiringRationalist) on Reverse Regulatory Capture · 2024-04-11T04:38:23.348Z · LW · GW

Do you have any examples (or likely examples) of this happening?

Comment by NoSignalNoNoise (AspiringRationalist) on What are the weirdest things a human may want for their own sake? · 2024-03-21T03:41:01.970Z · LW · GW

Spicy food. Plants evolved capsaicin production in order to deter mammals from eating them, yet many humans (myself included) like eating plants specifically because they contain capsaicin.

Comment by NoSignalNoNoise (AspiringRationalist) on Don't Endorse the Idea of Market Failure · 2024-03-03T04:23:50.759Z · LW · GW

This post reads as a call to action (or inaction), but it's not clear what you're saying people should do. Can you be more explicit about that?

Comment by NoSignalNoNoise (AspiringRationalist) on On "Geeks, MOPs, and Sociopaths" · 2024-01-20T20:11:11.076Z · LW · GW

Normal subcultures don't have infosec requirements, let alone infosec requirements effective enough for intelligence agencies

This link is broken

Comment by NoSignalNoNoise (AspiringRationalist) on How to (hopefully ethically) make money off of AGI · 2023-11-07T06:39:43.393Z · LW · GW

How would you recommend shorting long-dated bonds? My understanding is that both short selling and individual bond trading have pretty high fees for retail investors.

Comment by NoSignalNoNoise (AspiringRationalist) on Text Posts from the Kids Group: 2022 · 2023-09-28T03:20:16.704Z · LW · GW

Anna: Papa, I have a problem: [explains problem in detail]

Me: Is it a problem you can fix?

Anna: Yes! [Fixes problem]

I need to do a better job of asking myself this question.

Comment by NoSignalNoNoise (AspiringRationalist) on Is Bjorn Lomborg roughly right about climate change policy? · 2023-09-28T01:02:51.974Z · LW · GW

An important thing that this analysis leaves out is the uncertainty regarding feedback loops. If e.g. warming causes permafrost to melt and release more greenhouse gasses, there is a possibility of a runaway process that results in catastrophic warming. We don't know how bad the tail risks are, and an analysis that looks at the median case doesn't address that issue.

Comment by NoSignalNoNoise (AspiringRationalist) on What social science research do you want to see reanalyzed? · 2023-09-22T00:32:41.245Z · LW · GW

I'd like to know whether deep canvassing actually works.

Comment by NoSignalNoNoise (AspiringRationalist) on ACX Prediction Contest Discussion · 2023-01-07T05:01:08.556Z · LW · GW

The deadline to post your blind mode predictions is coming up on the 10th. This weekend is a great time to do them if you haven't yet.

Comment by NoSignalNoNoise (AspiringRationalist) on Noting an unsubstantiated communal belief about the FTX disaster · 2022-11-13T17:27:38.531Z · LW · GW

His 80000 interview suggests that he thought the chance of FTX blowing up is something between 1% and 10%. There he gives 50% odds for making more than 50 billion dollars that can be donated to EA causes.

If someone is saying that his action was negative in expectation, do they mean, that Sam Bankman-Fried lied about his expectations? Do they mean that a 10% chance of this happening should have been enough to tilt the expectation to be negative under the ethical assumptions of longtermism that puts most of the utility that's produced in the far future? Are you saying something else?

I wish I had any sort of trustworthy stats about the success rate of things in the reference class of steal from one pool of money in order to cover up losses in another pool of money, in the hope of making (and winning) big bets in the second pool of money to eventually make the first pool of money whole. I would expect the success rate to be very low (I would be extremely surprised if it were as high as 10%, somewhat surprised if it were as high as 1%), but it's also the sort of thing where if you do it successfully, probably nobody finds out.

Do Ponzi schemes ever become solvent again? What about insolvent businesses that are hiding their insolvency?

Comment by NoSignalNoNoise (AspiringRationalist) on Noting an unsubstantiated communal belief about the FTX disaster · 2022-11-13T17:19:59.356Z · LW · GW

I think approximately no one audits people's books before accepting money from them. It's one thing to refuse to accept money from a known criminal (or other type of undesirable), but if you insist that the people giving you money prove that they obtained it honestly, then they'll simply give that money to someone else instead.

Comment by NoSignalNoNoise (AspiringRationalist) on Noting an unsubstantiated communal belief about the FTX disaster · 2022-11-13T17:15:33.919Z · LW · GW

This is basically a Quirrell moment in real life; a massive proportion of people on LW are deferring their entire worldview to obvious supervillains.

Who are the obvious supervillains that they're deferring their entire worldview to? And who's deferring to them?

Comment by NoSignalNoNoise (AspiringRationalist) on Sabotage challenge · 2022-03-07T00:14:34.391Z · LW · GW

If in a job that's important for the war effort, be a stickler for following all rules and official procedures. Escalate decisions so that things don't get done without official input from higher-ups.

Comment by NoSignalNoNoise (AspiringRationalist) on Are our community grouphouses typically rented, or owned? · 2022-03-02T05:24:37.186Z · LW · GW

The majority rent (I've lived in a few, all of which, including the one I live in now, rented).

I believe the main reasons for this are:

  • People who live in rationalist group houses are disproportionately young and live in expensive areas, which makes it hard to buy a house,
  • There's a lot of variability in how long people live in rationalist group houses, and
  • Figuring out the ownership structure is complicated.

The first point is fairly self-explanatory, but I'll say a bit more about the other two.

There are several sorts of people who choose to live in a rationalist group house:

  1. People who would rather live in a rationalist group house than live alone or just with a partner/family,
  2. People who want to live in a rationalist group house until they find a partner and settle down,
  3. People who thought they wanted to live in a rationalist group house but decided it wasn't for them (often because they find out they're more introverted than they realized or want more control over their living space than a group house offers),
  4. People who can't afford to live alone so they live in a group house, and given that they need to live with other people, they'd prefer rationalists, and
  5. People who are moving to or explicitly temporarily living in a particular city (e.g. to study) who want their housing to come with a rationalist-type social circle,
  6. Partners of rationalists who themselves aren't rationalists, and
  7. People in the rationalist community who can live in a rationalist house more cheaply or more conveniently than somewhere else (often but not always because a room is temporarily vacant).

Most of these kinds of people aren't going to stick around very long. That's fine; the temporary (a few months to a year) residents of the rationalist group houses I've lived in have generally been positive additions to the house, so I wouldn't want to exclude them.

Because most of the people who might want to live in a rationalist house won't be sticking around that long, it doesn't make sense for everyone to own it. Which brings us to the question of some subset of the residents owning the house.

Last year, a friend and I looked into buying a house together to turn into a group house (where we would rent the rooms out to other residents). Things I learned from that process were (I expect this to vary a lot by geography, and I know very little about New Zealand's housing market):

  • It can be hard to find something that matches multiple people's constraints (in terms of price, location, size, features).
  • Co-owning a house with someone (other than a spouse) is legally complicated and requires a good contract and a competent lawyer. Especially if there's also a mortgage involved.
  • Touring houses is a lot of work.
  • Most houses for sale have a lot wrong with them and the permitting process for transforming them to be the way you want is slow and unpredictable.
  • Figuring out whether there'd be sufficient interest in a rationalist group house in a particular location is hard.

There were a couple houses that we came close to want to make an offer on (though we still hadn't figured out the legal issues around co-ownership). Then my friend accepted a job offer in another city, which ended that project.

None of this means you shouldn't buy a house for this purpose under the right circumstances. I think those circumstances are:

  • Someone in the group has the ability to buy such a house.
  • Enough people are interested and have sufficiently legible requirements regarding price, location, size, and amenities.
  • The prospective buyer is ok with the house ending up not being a rationalist group house (and either living in themself not as a group house or turning it into a regular rental property).
Comment by NoSignalNoNoise (AspiringRationalist) on Omicron Post #14 · 2022-01-15T23:54:42.969Z · LW · GW

There's also an asymmetry between gains and losses, partly due to prospect theory, and partly due to decreasing marginal utility. I bet a lot of people would answer differently if they were asked what they would choose if given the choice between receiving the money vs. going back to the way things were before.

Comment by NoSignalNoNoise (AspiringRationalist) on Omicron variolation? · 2022-01-08T03:57:34.831Z · LW · GW

I think it depends on whether you think there will be an omicron booster by the time the next variant comes along. If there is, you'll have gotten Covid for nothing.

Comment by NoSignalNoNoise (AspiringRationalist) on Omicron Post #11 · 2021-12-31T19:57:52.086Z · LW · GW

Abbot flat out denies the FDA’s claim of potential lowered test sensitivity, says their tests are as effective against Omicron as they were against previous variants.

The link here appears to be a mattress ad.

Comment by NoSignalNoNoise (AspiringRationalist) on COVID Skepticism Isn't About Science · 2021-12-29T22:25:06.490Z · LW · GW

How often do people talk about tradeoffs between multiple sacred values?

Comment by NoSignalNoNoise (AspiringRationalist) on COVID Skepticism Isn't About Science · 2021-12-29T20:21:18.262Z · LW · GW

This story makes sense for describing how people might believe conspiracy theories because they oppose lockdowns, but I don't think a similar story would apply for opposition to vaccines. Following this line of thinking, I think the sequence of events is:

  1. Disease breaks out.
  2. Public health authorities respond to the disease with high-cost preventative measures.
  3. People respond to those preventative measures by becoming hostile to public health measures.
  4. People's hostility to public health measures oppose vaccines even though they're much lower cost and much more effective than the measures that led to them becoming hostile to public health measures in the first place.
Comment by NoSignalNoNoise (AspiringRationalist) on COVID Skepticism Isn't About Science · 2021-12-29T20:12:51.891Z · LW · GW

An important aspect of this is that it involves a tradeoff between a sacred value (preventing death) and a secular value (avoiding restrictions). When it's not socially acceptable to have a frank discussion of the real costs and benefits of various restrictions, it becomes easier for people who oppose the restrictions to pretend that the benefits of the restrictions don't exist (aka the disease isn't real or isn't serious).

Comment by NoSignalNoNoise (AspiringRationalist) on Omicron: My Current Model · 2021-12-28T19:40:04.261Z · LW · GW

Vaccination with three doses is protective against infection by Omicron, but less protective than vaccines were against Delta. As a rule of thumb I am currently acting as if a booster shot is something like 60%-70% protective against infection but I don’t have confidence in that number. The main protection is still against severe disease, hospitalization and death.

Two questions about this:

  1. Do you mean that a booster is 60-70% effective relative to being "fully" vaccinated but not boosted, or do you mean that being boosted is 60-70% effective relative to being unvaccinated?
  2. How did you reach this conclusion? Based on the Pfizer press release, I had been treating being boosted as 95% effective (relative to being unvaccinated) since the level of neutralizing antibodies against Omicron with 3 doses was the same as the level against original Covid with 2 doses, and 2 doses were 95% effective against original.
Comment by NoSignalNoNoise (AspiringRationalist) on Solstice Quaker Darkness Circling · 2021-12-28T14:46:21.043Z · LW · GW

What was the weather like? I like this idea, and I wonder how well it would work under different weather conditions.

Comment by NoSignalNoNoise (AspiringRationalist) on Omicron Post #4 · 2021-12-07T01:48:34.668Z · LW · GW

Either it’s bad or weird data, or Omicron somehow puts a ton more virus into the wastewater, and then there’s nothing to see here.

If Omicron somehow puts a ton more virus into the wastewater, that tells us something interesting about the virus. Maybe it somehow infects the digestive tract more effectively than other strains.

Comment by NoSignalNoNoise (AspiringRationalist) on Covid 11/18: Paxlovid Remains Illegal · 2021-11-22T04:28:55.843Z · LW · GW

How should we determine when there's adequate supply? I imagine calling pharmacies and asking "if, hypothetically, I got Covid, would everyone in my household be able to get Paxlovid?" would work very well.

Comment by NoSignalNoNoise (AspiringRationalist) on Is there a clearly laid-out write-up of the case to drop Covid precautions? · 2021-11-11T02:51:04.089Z · LW · GW

Facebook discussion on this.

Comment by NoSignalNoNoise (AspiringRationalist) on Let's Go Back To Normal · 2021-11-10T02:14:04.659Z · LW · GW

Vietnam

Comment by NoSignalNoNoise (AspiringRationalist) on An Unexpected Victory: Container Stacking at the Port of Long Beach · 2021-10-31T14:22:26.357Z · LW · GW

If there are specific problems with those regulations, shouldn't a legislator representing a district with a lot of forestry or ranching be able to propose a sensible solution with little opposition?

Comment by NoSignalNoNoise (AspiringRationalist) on An Unexpected Victory: Container Stacking at the Port of Long Beach · 2021-10-31T14:16:15.039Z · LW · GW

The default approach is to try to get the attention of the highest-ranking person they can think of, but this runs afoul of the exact mechanism you mention where attention is precious and the higher the rank, the more fierce the competition for it, and the higher the threshold we need to reach to direct them. But I think this is a power-law distribution, which is to say that as you go down the ladder of hierarchy the attention threshold drops rapidly.

To sum up, we can mitigate the attention problem by aiming as low on the totem pole as possible, and providing as explicit an action as possible.

I wonder if, in this context, that would have meant trying to get the attention of Transportation Secretary Buttigieg rather than President Biden. Or the mayor of Long Beach.

Comment by NoSignalNoNoise (AspiringRationalist) on Come Build Affordable Housing! · 2021-08-01T20:13:16.633Z · LW · GW

Right, but that impacts whether it's actually profitable to build them.

Comment by NoSignalNoNoise (AspiringRationalist) on Come Build Affordable Housing! · 2021-07-26T23:25:23.025Z · LW · GW

If you buy one, l assume you can't then rent it out at market rate? What restrictions are there on your ability to resell it? I would expect that to massively decrease these units' value to potential buyers.

Comment by NoSignalNoNoise (AspiringRationalist) on The value of low-conscientiousness people on teams · 2021-06-20T20:19:22.112Z · LW · GW

As long as we're going off on tangents, does anyone know a name for the bias where Oxonians look like they're doing things effortlessly?

I suspect the following is a common psychological failure mode, and I want a term to refer to it:

  1. See someone doing something amazing and making it look easy
  2. Try to do something similar (or imagine trying to)
  3. Realize (or assume) that it's hard and will take a lot of work
  4. Conclude that because it's easy for the other person and hard for you, you must be bad at it (when actually it's hard for the other person too, but you just don't see the work that they put into it)
  5. Since you've concluded that it's hard and you're bad at it, you give up
Comment by NoSignalNoNoise (AspiringRationalist) on Are bread crusts healthier? · 2021-06-19T01:43:01.260Z · LW · GW

The intuition might also come from whole grains generally being healthier and darker than refined grains. A naive attempt to generalize that might conclude that the darkest part of the bread is the healthiest.

Comment by NoSignalNoNoise (AspiringRationalist) on What topics are on Dath Ilan's civics exam? · 2021-04-28T22:14:09.321Z · LW · GW

Given the following charts, statistics, and arguments based on those charts and statistics, point out the important flaws in the arguments and state what unjustified conclusions the arguers are trying to cause you to reach.

Comment by NoSignalNoNoise (AspiringRationalist) on The Fall of Rome, II: Energy Problems? · 2021-04-23T21:39:47.559Z · LW · GW

For wood and charcoal to be expensive, forests don't necessarily need to be depleted. Instead, it can be due to higher transportation costs (wood and charcoal are heavy). As the empire became less secure during and after the crisis of the 3rd century, I would expect long-distance transportation to have become less safe (due to banditry) and therefore more expensive.

Comment by NoSignalNoNoise (AspiringRationalist) on Covid 4/22: Crisis in India · 2021-04-23T04:02:17.177Z · LW · GW

The important question about Alaska opening up vaccines to tourists is whether non-Americans will be able to get in.

Comment by NoSignalNoNoise (AspiringRationalist) on Speculations Concerning the First Free-ish Prediction Market · 2021-04-01T00:41:37.855Z · LW · GW

Regarding betting on inflation, TIPS already exist.

Comment by NoSignalNoNoise (AspiringRationalist) on "You and Your Research" – Hamming Watch/Discuss Party · 2021-03-21T15:58:32.652Z · LW · GW

Is the idea to watch it when the event starts or to watch it beforehand?