Comment by lifelonglearner on How much background technical knowledge do LW readers have? · 2019-07-12T05:46:16.398Z · score: 4 (2 votes) · LW · GW

FWIW, both linear algebra and multivariable calculus are required for students at UC San Diego, which is a large public institution.

(Although it's a little tricky as our university has sub-colleges, not all of which require both).

Comment by lifelonglearner on Can I automatically cross-post to LW via RSS? · 2019-07-08T05:52:04.505Z · score: 3 (2 votes) · LW · GW

Thanks! I'm a little worried about my own signal-to-noise ratio for now, so I'll do that later on if I feel my short-form quality improves.

Comment by lifelonglearner on Can I automatically cross-post to LW via RSS? · 2019-07-08T05:51:34.135Z · score: 2 (1 votes) · LW · GW

Meant for this to be a reply.

Can I automatically cross-post to LW via RSS?

2019-07-08T05:04:55.829Z · score: 10 (3 votes)
Comment by lifelonglearner on MLU: New Blog! · 2019-06-13T18:09:19.386Z · score: 3 (2 votes) · LW · GW

Yeah, Netlify was really easy to set up, so I recommend them!

I don't think I'll be able to migrate blog comments because I'm still not sure what I'll do for comments on the new blog, actually.

I don't want to use Disqus because it's bulky, other options cost money, so maybe a self-hosted option...

Comment by lifelonglearner on MLU: New Blog! · 2019-06-13T02:27:04.144Z · score: 2 (1 votes) · LW · GW

Long-term, I'll probably do some more organizing to put posts into sequences / other things to improve readability.

Right now, the focus is on updating/editing old posts, so that the main page is a set of polished essays that work as standalones.

MLU: New Blog!

2019-06-12T04:20:37.499Z · score: 18 (5 votes)
Comment by lifelonglearner on Go Do Something · 2019-05-21T15:47:12.426Z · score: 19 (6 votes) · LW · GW

This seems reasonable.

See also Yudkowsky in Inadequate Equilibria for a similar sentiment:

Try to spend most of your time thinking about the object level. If you’re spending more of your time thinking about your own reasoning ability and competence than you spend thinking about Japan’s interest rates and NGDP, or competing omega-6 vs. omega-3 metabolic pathways, you’re taking your eye off the ball.
Comment by lifelonglearner on Why books don't work · 2019-05-12T05:47:10.095Z · score: 8 (4 votes) · LW · GW

I don't think the author disagrees all that much with you. I'm reading his claim as something more like "the default attitude (some) people have towards reading does not set them up for good learning".

In the essay, he acknowledges the role that effort and metacognition play in making the actual learning happen. The actionable parts I found useful were at the end where he was hypothesizing about improved mediums, e.g. an online textbook with spaced repetition built in to facilitate recall.

Comment by lifelonglearner on Why books don't work · 2019-05-12T05:45:08.217Z · score: 9 (3 votes) · LW · GW

I think I agree that the generalizations you cited were hasty, especially as there is no formal review on those matters. I, too, find that I can get value out of books and can probably do a better job than just giving a brief summary.

Let me try to pick out the parts that I liked and see if you disagree with those:

The main part of the thesis that I found interesting was the analogy between books and lectures. In both cases, there is the potential of transmissionism as one naive way of thinking about how information gets absorbed.

It's clearer, however, that lectures in an educational setting aren't enough. Without problem sets, examples, and other applications, most of the "real learning" doesn't happen. (i.e. "mathematics is not a spectator sport") Good professors seem to recognize this and supplement their teachings accordingly.

Then, the author points out that a similar acknowledgment for books is not the norm. There is still room for improving the medium, and one example he gave was the spaced repetition enhanced online book. There is an undercurrent of "you as the reader need to put in effort to get value out of your reading", which I agree with. It's a different state of mind when I'm reading for insight vs reading for fun. In the first case, I might re-read passages, skip around, stop for a bit, take some notes, etc. etc. whereas in the second case, I'm probably just reading rather quickly from cover to cover.

Why books don't work

2019-05-11T20:40:27.593Z · score: 16 (11 votes)
Comment by lifelonglearner on Crypto quant trading: Naive Bayes · 2019-05-08T20:20:50.762Z · score: 2 (1 votes) · LW · GW

I see. Thanks for providing the additional info!

Comment by lifelonglearner on Crypto quant trading: Naive Bayes · 2019-05-08T04:50:39.359Z · score: 6 (3 votes) · LW · GW

I've read over briefly both this article and the previous one in the series. Thank you for putting these together!

What I'm curious about in quant trading is the actual implementation. Once you, say, have a model which you think works, how important is latency? How do you make decisions about when to buy / sell? (Partially echoing Romeo's sentiment about curiosity around stop losses and the actual nitty-gritty of extracting value after you think you've figured something out.)

345M version GPT-2 released

2019-05-05T02:49:48.693Z · score: 30 (11 votes)
Comment by lifelonglearner on Neural networks for games · 2019-05-02T04:53:41.463Z · score: 3 (2 votes) · LW · GW

Heads-up: nowadays, when people talk about neural networks for games, they really mean deep learning combined with reinforcement learning.

Back to your question: When you don't have a log of games, you typically have some other way of assessing performance, e.g. assigning a "score" to the state of the game, which you can quantify and optimize.

For a specific well-known example, I think this paper on training to play Atari games with deep reinforcement learning goes over a lot of the actual math / implementation details.

Moving to a World Beyond “p < 0.05”

2019-04-19T23:09:58.886Z · score: 25 (10 votes)
Comment by lifelonglearner on AI Safety Prerequisites Course: Basic abstract representations of computation · 2019-03-14T17:06:07.770Z · score: 9 (4 votes) · LW · GW

I've looked a little bit at the RAISE website, and I've looked at the overview of curriculum topics, and I'm finding it a little...sparse, maybe? (I haven't actually looked at the class materials on grasple though, so maybe there's more stuff there.) I'm wondering how realistic it would be for someone to start engaging with MIRI-esque topics after learning just the courses RAISE has outlined.

At least for the prerequisites course, these are all topics covered throughout the first two years of a typical undergraduate computer science degree. And that doesn't seem like quite enough.

EX: TurnTrout's sequence of essays on their journey to become able to contribute towards MIRI-esque topics seems to span a much greater gamut of topics (linear algebra, analysis, etc.) at greater depth, closer to what one might cover in graduate school.

I guess, to operationalize, I'm curious about:

1. What target audience RAISE has in mind (technical people looking for a refresher, people who have had zero real exposure to technical subjects before, etc. etc.) for their materials.

2. What degree of competence RAISE expects people to come out of the curriculum with, either best-case or average-case.

3. In the best case, how many units of material do you think RAISE can product? In other words, is it enough for students to study RAISE's material for a 6-month long curriculum? 1 year long?


(Of course, it's also much easier from my position to be engaging/critiquing existing works, than to actually put in the effort to make all of this happen. I don't mean any of the above as an indictment. It's admirable and impressive that y'all have coordinated to make this happen at all!)

Pedagogy as Struggle

2019-02-16T02:12:03.665Z · score: 14 (6 votes)

Doing Despite Disliking: Self‐regulatory Strategies in Everyday Aversive Activities

2019-01-19T00:27:05.605Z · score: 14 (3 votes)

mindlevelup 3 Year Review

2019-01-09T06:36:01.090Z · score: 19 (5 votes)
Comment by lifelonglearner on What exercises go best with 3 blue 1 brown's Linear Algebra videos? · 2019-01-01T22:17:50.007Z · score: 15 (6 votes) · LW · GW

If you want a proof-based approach, Linear Algebra Done Right is the typical go-to that's also on the MIRI page. I went through maybe the first 3/4ths of it, and I thought it was pretty good, in terms of number of exercises and helping you think about manipulating vector spaces, etc. in a more abstract sense.

Otherwise, I've heard good things about Gilbert Strang's MIT OCW course here:

In general, I think that 3B1B's videos are really good for building intuition about a concept, but trying to do exercises off of the pedagogy in his videos alone can be quite challenging, especially as he often assumes some mastery with the subject already. (EX: In the eigen-stuffs video, he doesn't actually explain how to find the eigenvalues of a matrix.)

Thus, I think it makes more sense to stick to a traditional textbook / course for learning linear algebra and using 3B1B as supplementary stuff for when you want a visual / different way of looking at a concept.

Also, it might be worth checking in to see what you want to learn linear algebra for. I suspect there are more domain specific resources if, for example, you cared about just the useful parts of linear algebra used in machine learning (dimensionality reduction, etc.).

Comment by lifelonglearner on How can I most effectively improve my writing abilities? · 2019-01-01T17:20:11.137Z · score: 8 (5 votes) · LW · GW

I think the overarching thing to do is to simply write more. To that end, there are a lot of ways to make this happen. For example: participating in NaNoWriMo, committing to writing a sequence of articles about topic X, engaging in discourse online, summarizing research papers for a larger audience, writing guest articles for a blog, or journaling.

As for improving your actual workflow, I think that the Typical Writing Class you take in school largely gets it right. For example, writing outlines / summaries before starting the actual writing, asking people to proofread, have several drafts / edit, and try to imagine reading it from the reader's perspective w/o the extra context you have as a writer.

(I think most of the dissatisfaction I had with the skills taught to me during class had a lot more to do with the context of "Oh man, I have to write this thing using this technique for school and not of my own volition?" rather than the skills themselves not being very good.)

Comment by lifelonglearner on Card Collection and Ownership · 2018-12-27T17:18:50.396Z · score: 2 (1 votes) · LW · GW

Do you have any thoughts written up anywhere on what a game you'd design would look like? I've been following all of your MTG / digital CCG articles with a lot of interest; I'd be really curious which aspects of game design you'd like to incorporate.

Comment by lifelonglearner on What self-help has helped you? · 2018-12-21T16:31:39.843Z · score: 5 (3 votes) · LW · GW

I am wondering if there are commonalities between what different self-help things are doing. For example, it seems that a lot of self-help is focused on changing our default actions, ala debiasing, so there is a train of thought that starts with cognitive biases and goes from there.

Comment by lifelonglearner on What self-help has helped you? · 2018-12-21T02:47:08.161Z · score: 4 (3 votes) · LW · GW

A related question I'm wondering about, which seems related to this is "Why does self-help work? What is it doing?"

Not 100% related to the question, but maybe the discussion here in the comments could spark more (or another question).

Comment by lifelonglearner on Letting Others Be Vulnerable · 2018-11-21T17:17:16.175Z · score: 3 (2 votes) · LW · GW

Oh, right, yes, I didn't think about that.

That is a good example where the above advice I listed, as well as the following models, break.

Comment by lifelonglearner on Letting Others Be Vulnerable · 2018-11-21T04:44:04.647Z · score: 2 (1 votes) · LW · GW

Oh, right. I think an implicit thing here is "for people you want to be close to, this makes sense to do".

In other cases, as I sorta skimmed over, having simplified models, relying on norms/roles, etc. etc. is usually enough to get by.

Letting Others Be Vulnerable

2018-11-19T02:59:21.423Z · score: 32 (16 votes)
Comment by lifelonglearner on Alignment Newsletter #30 · 2018-10-30T02:48:07.128Z · score: 7 (4 votes) · LW · GW

The Learning from Untrusted Data paper was really clear, and I found the results surprising. Thanks for sharing and for consistently putting this out week after week. I really like the combination of the summary + your personal thoughts.

Comment by lifelonglearner on What will the long-term future of employment look like? · 2018-10-24T22:51:25.691Z · score: 4 (2 votes) · LW · GW

Specifically for the example of social influencers, I wonder if roles like this might be limited in capacity because, by virtue of their role, they require a large following. Given the way popularity and status tends to be non-uniformly distributed across social groups, it might not be a very large role for most people to fit into.

But overall, I think the underlying theme of humans-as-valuable-because-they-are-human and the examples to Etsy charging a premium on human-made things hits the nail on the head as for where residual value might lie in an automated future.

I suppose one corollary, too, is that it will be human desires which comprise the demand for these goods. If it turns out that automation also shifts the demand for certain services, via machines sending orders to machines, then this might also limit future options for humans?

Comment by lifelonglearner on LW Update 2018-10-01 – Private Messaging Works · 2018-10-02T04:41:06.023Z · score: 4 (2 votes) · LW · GW

Bug report: I can no longer view old messages.

I clicked on my profile, went to private messages, and then clicked on something from the conversations sidebar.

Every message had the following red error:

"Error: TypeError: Cannot read property '_id' of null"

Comment by lifelonglearner on Thoughts on tackling blindspots · 2018-09-27T14:50:37.150Z · score: 10 (6 votes) · LW · GW

You might already be aware, but there's the Unofficial CFAR Canon List which compiles a lot of their earlier material (though some things have now changed) that someone put together a while back.

If you're looking for more derivative content written by people who have gone to CFAR workshops:

There's also the Hammertime sequence alkjash wrote and the Instrumental Rationality sequence I wrote.

Comment by lifelonglearner on Owen's short-form blog · 2018-09-16T04:27:02.248Z · score: 2 (1 votes) · LW · GW

Yup! I just set up one up here. I just grabbed an off-the-shelf Python package so let me know if there are any issues.

Owen's short-form blog

2018-09-15T20:13:37.047Z · score: 13 (6 votes)

Communication: A Simple Multi-Stage Model

2018-09-15T20:12:16.134Z · score: 13 (4 votes)
Comment by lifelonglearner on Changing main content font to Valkyrie? · 2018-08-25T02:02:20.917Z · score: 3 (2 votes) · LW · GW

Note: I am on Windows and find the font fine. I am wondering how many people on Windows find it hard to read?

Comment by lifelonglearner on Rationalist Community Hub in Moscow: 3 Years Retrospective · 2018-08-25T02:00:18.314Z · score: 9 (6 votes) · LW · GW

Thanks for giving us this English overview! I was very curious when someone mentioned Kocherga in a comment a while back, but the site was in Russian and not very clear.

This is really cool to read about!

Comment by lifelonglearner on Is there a practitioner's guide for rationality? · 2018-08-13T17:26:52.582Z · score: 6 (4 votes) · LW · GW

Ikaxas has already linked to alkjash's sequence, which is excellent. I also wrote a sequence, mostly on habits and planning here.

Comment by lifelonglearner on A Rationalist's Guide to... · 2018-08-09T23:09:38.266Z · score: 10 (3 votes) · LW · GW

I think this is largely correct and points at where some of the larger bottlenecks are.

It's not about finding a list of good resources. There are a lot of those already. It's about what happens next. Things like:

  • Getting yourself to actually read said resources.
  • Figuring out ways of making the material stick.
  • Looking for applications, tracking your progress.
  • Repeating all of the above, over and over.
Comment by lifelonglearner on Open Thread August 2018 · 2018-08-09T21:11:02.231Z · score: 2 (1 votes) · LW · GW

Also interested!

Comment by lifelonglearner on Hammertime Day 8: Sunk Cost Faith · 2018-08-04T21:59:16.071Z · score: 5 (3 votes) · LW · GW

The original author had them removed. You can find quite a few of them now on Medium instead, here:

Fading Novelty

2018-07-25T21:36:06.303Z · score: 24 (14 votes)
Comment by lifelonglearner on Generating vs Recognizing · 2018-07-14T14:29:04.586Z · score: 2 (1 votes) · LW · GW

Oh, right, yup. The P vs NP analog is a very good parallel.

Generating vs Recognizing

2018-07-14T05:10:22.112Z · score: 16 (6 votes)
Comment by lifelonglearner on The Craft And The Codex · 2018-07-10T00:53:36.299Z · score: 15 (4 votes) · LW · GW

Here's a comment by Montfort that I liked from the SSC comments:

I can sort of buy the idea of the comments as a dojo – but there’s no personalized instruction or membership fees or belts or anything. The learning here is extremely self-directed, success is hard to judge objectively, and commenters come and go all the time. Still, even just an empty building with some mats where people can show up and practice is something.

That is, if you wanted to design a place to practice “rationality” skills from the ground up, I’m not sure it would look like this. But the blog and comment section can serve multiple purposes at once.

I have competing intuitions about just how much structure and scaffolding is useful for practice. I don't think schools do a lot of things right, but at least here in the US, we can count on them getting people to be literate and able to do basic arithmetic. I'm wondering what a "minimum set" of rationality techniques would look like, things you could teach to people at scale.

Comment by lifelonglearner on RAISE is looking for full-time content developers · 2018-07-10T00:48:57.741Z · score: 10 (5 votes) · LW · GW

It's actually European convention vs US:

Comment by lifelonglearner on Problem Solving with Mazes and Crayon · 2018-06-24T13:50:24.318Z · score: 6 (1 votes) · LW · GW

I really like the maze analogy! The pictures help a whole bunch in explaining the concepts!

Related: I don't know if this has a direct analog to a CS algorithm, but one other heuristic which seemed like it might have been good to try was drawing a line from both start and finish incrementally to see how they could touch in the middle.

Do Conversations Often Circle Back To The Same Topic?

2018-05-24T03:07:38.516Z · score: 9 (2 votes)
Comment by lifelonglearner on Ikaxas' Shortform Feed · 2018-05-18T21:42:13.879Z · score: 5 (1 votes) · LW · GW

This feels, at least surface-level, similar to what I was trying to get at here about how things can be self-defeating. Do you also think the connection is there?

Comment by lifelonglearner on [CKC] [May 2018] What subjects are most important for an AI safety researcher to know? (Open Call) · 2018-05-18T21:32:52.675Z · score: 18 (4 votes) · LW · GW

I think Raemon has a good overview of everything here:

In addition, MIRI has a recommended textbooks list

Also, CHAI has a textbooks list here

Also, Toon Alfrink has started RAISE to help bring people up to speed.

(META: I think this topic has a very long history of already being discussed. See, for example, lukeprog on the best textbooks. Also, we've tried having monthly open-thread-like-discussion-things in the past, and they sort of flopped. So just a heads-up if you're aiming for something similar.)

Comment by lifelonglearner on Instrumental Rationality 6: Attractor Theory · 2018-05-17T18:45:07.457Z · score: 6 (2 votes) · LW · GW

I think they're pretty different. My read on on Lotus Eating is basically "addictiveness is considered harmful and there's a bunch of things out there that can hijack your attention".

My thoughts on Attractor Theory are more about "note the ways your preferences will change in response to actions you take and act accordingly". While this certainly often includes how certain activities can be spirals, I think it goes more breadth-wise and prescribes a more general strategy for which actions to take.

Musings on Lotus Eating can get a lot deeper into the "whys" and "hows" of whether or not X counts as Lotus Eating, whether or not you could secretly benefit from X, etc. Attractor Theory just notes that getting spiraled in could be a consequence of certain actions, and you can take this into account.

Comment by lifelonglearner on Instrumental Rationality 4.3: Breaking Habits and Conclusion · 2018-05-17T18:42:01.031Z · score: 6 (2 votes) · LW · GW


Alas, for cue independent habits, I'm not quite sure what would be effective. (The literature doesn't cover these well.) For a probably ineffective start, maybe consider intentionally substituting something else at the times where you notice yourself doing the thing the most?

Comment by lifelonglearner on Personal relationships with goodness · 2018-05-14T19:33:12.849Z · score: 10 (2 votes) · LW · GW

I don't think I have too much to add, but I thought it was nice to see these different takes fleshed out.

My own views are a mix of your last two. It goes a little like:

"At best, humans are operating off leaky abstractions when it comes to 'doing good'. It's not realistic to expect yourself to do The Goodest Thing (TM) all the time, even when it seems like it's doable. Just like how you can't always swim to the other side of the shore, even if you can see it from across the waves. There's a lot of other factors involved, and doing things like taking breaks are instrumental in the long run to avoid burnout."

Comment by lifelonglearner on Turning 30 · 2018-05-08T19:04:32.644Z · score: 14 (3 votes) · LW · GW

I really enjoyed reading this personal piece. 10 years is a long time to span, and I found myself wondering what other adventures happened, but I guess that might have just been compression / focusing on the important bits.

Comment by lifelonglearner on Advocating for factual advocacy · 2018-05-06T17:01:18.467Z · score: 10 (3 votes) · LW · GW

I think that providing people with additional actions / alternatives seems good, especially if you can then feel better about yourself because it ended up being easier than you thought to act on a certain moral view.

I think this might be valid mainly places where there do exist easy ways to switch / make the action easier. EX: Providing public transportation credits for commuters, or redirecting their existing charity payments to another cause.

I think the right question to frame it as, then, is something like "How much additional work are we asking of you to do X, and how can we make it easier for you to take X?"

Comment by lifelonglearner on A possible solution to the Fermi Paradox · 2018-05-05T15:55:06.828Z · score: 5 (1 votes) · LW · GW

Quick clarification: Does the reasoning in the 99:1 scenario differ from the sort of reasoning used in anthropic reasoning? (I'm unsure of how that differs a lot from survivorship bias.)

Comment by lifelonglearner on Program Search and Incomplete Understanding · 2018-05-03T21:53:54.811Z · score: 6 (2 votes) · LW · GW

This was a very neat writeup of the ways that progress on problems can move. I didn't have a good sense of these being distinct, or of denominating between different types of understanding. Thanks for putting this together!

Fyi the Levin link seems to be broken.

Comment by lifelonglearner on Ikaxas' Hammertime Final Exam · 2018-05-01T03:50:21.789Z · score: 12 (3 votes) · LW · GW

Hurrah for taking the exam!

For my own views, I think it's often the case that people bias towards not changing their mind. This feels like part of the impetus for the push-back early on in stressing the benefit in changing your mind. (Duncan Sabien has a neat simple analogy of a pendulum getting counterbalanced back and forth on this).

However, I don't think I've even really figured out the trick of changing my mind yet. There have been a few instances recently where nothing felt very sacred, and that sort of fluidity felt like it could be a powerful asset when trying to updated towards certain things in reality.

Comment by lifelonglearner on Meditations on the Medium · 2018-04-30T20:50:11.154Z · score: 5 (1 votes) · LW · GW
Modulo the motivating problem you mention, the digital textbook medium you propose seems to solve all three, assuming new, revised "editions" are released every now and then. Do you have any thoughts about how it compares with a Wiki medium?

Hm. I think that Wikis might be easier to maintain, perhaps? There seem to be a lot of additional considerations when you're writing to teach and not just to explain / inform.

5) Explaining ideas in detail makes it harder to follow the big picture.
On re-readings of the material, sometimes I wish I could "compress" explanations I already understand to focus on the higher picture and relations between ideas. Skimming seems the thing to do, but it requires some sort of indexing, and right now I feel like digital media should be able to offer more powerful ways of doing this.

This matches my experience. I'd like some way of "collapsing" certain paragraphs, or labeling them as "basically just about X" to help with overall comprehension.

Meditations on the Medium

2018-04-29T02:21:35.595Z · score: 46 (12 votes)
Comment by lifelonglearner on Noticing the Taste of Lotus · 2018-04-29T00:38:02.477Z · score: 5 (1 votes) · LW · GW

I agree that this is something you can learn to notice and avoid. I find myself a little allergic to these things such that gamification actually feels repulsive. I tried out Pokemon Go recently, and it was terrifying how the game was set up, such that all your actions gave you little boosts, and the tapping led to bubbles, which burst and gave you colored points, etc.

There's something very much about being "in the thrall" of one of these traps, where you can get sucked in. I think the "one more X" mentality also captures a good part of this (where X is food, porn, videos, games, etc.)

Comment by lifelonglearner on The 3% Incline ( · 2018-04-27T02:03:21.441Z · score: 5 (1 votes) · LW · GW

Thanks for sharing this! I've been trying to collect thoughts on the nature of self-improvement for a while now. I think there's a lot here that has to with recalibrating our expectations, and the Typical Self Help Narrative certainly doesn't help.

In my head, I'm not quite sure what a 3% change looks like, but I'm also not sure what a better framing would be, especially one that allows you to stick with the compound interest analogy. Some potential contenders for alternate mental images:

1. Some sort of internal gamified EXP meter, borrowing from video games ("It's all about doing it over and over to rack up those points! Yeah, let me floss again today to rack up those Virtue Points(C)!")

2. Increased frequency, improved long-term performance. ("You might not do better today, but most days will become better!"), e.g. From journaling every 10 days to journaling every 5 days.

3. Base rates: It's 2 months for a habit. This is Empirically Backed. ("If you haven't stuck with X for 2 months, shut up and stick it out. Then we'll talk.")

Comment by lifelonglearner on The First Rung: Insights from 'Linear Algebra Done Right' · 2018-04-23T04:50:24.730Z · score: 10 (2 votes) · LW · GW


I just started going through this textbook! I'm working my through the proofs in chapter 1 right now, and I expect to progress through it as a supplement to the linear algebra course I'm taking right now.

Comment by lifelonglearner on Reducing Agents: When abstractions break · 2018-03-31T04:00:44.946Z · score: 9 (2 votes) · LW · GW

This was great, and I found it to be a clearer way of pointing at how typical "straw rationalist-y" ways of looking at decision-making might not work well in certain situations. I'm looking forward to the rest of the sequence!

Charting Deaths: Reality vs Reported

2018-03-30T00:50:00.314Z · score: 38 (11 votes)
Comment by lifelonglearner on Taking the Hammertime Final Exam · 2018-03-23T02:46:40.750Z · score: 6 (2 votes) · LW · GW

The thing I'd like to point to in 1) is "sometimes, you just run with the meta-knowledge that your brain is corrupted <right now>, and you just default to your Shoulds". And if any part of you tries to object, you silence it with a "Can't listen to you right now! You're compromised!"

Something like that has been useful for me in certain...sensitive, shall we say, situations.

<insert obvious caveats and problems with this approach here.>

Taking the Hammertime Final Exam

2018-03-22T17:22:17.964Z · score: 41 (11 votes)

A Developmental Framework for Rationality

2018-03-13T01:36:27.492Z · score: 61 (19 votes)

ESPR 2018 Applications Are Open!

2018-03-12T00:02:26.774Z · score: 4 (1 votes)

ESPR 2018 Applications Are Open

2018-03-11T20:07:45.460Z · score: 24 (5 votes)

Kegan and Cultivating Compassion

2018-03-11T01:32:31.217Z · score: 49 (12 votes)

Unconscious Competence and Counter-Incentives

2018-03-10T06:38:34.057Z · score: 37 (9 votes)

If rationality skills were Harry Potter spells...

2018-03-09T15:36:11.130Z · score: 67 (18 votes)

Replace Stereotypes With Experiences

2018-01-29T00:07:15.056Z · score: 16 (5 votes)

mindlevelup: 2 Years of Blogging

2018-01-06T06:10:52.022Z · score: 4 (1 votes)

Conceptual Similarity Does Not Imply Actionable Similarity

2017-12-30T05:06:04.556Z · score: 19 (9 votes)

Unofficial ESPR Post-mortem

2017-10-25T02:05:05.416Z · score: 58 (20 votes)

Instrumental Rationality: Postmortem

2017-10-21T06:23:31.707Z · score: 38 (11 votes)

Instrumental Rationality 7: Closing Disclaimer

2017-10-21T06:03:19.714Z · score: 13 (4 votes)

Instrumental Rationality 6: Attractor Theory

2017-10-18T03:54:28.211Z · score: 22 (9 votes)

Instrumental Rationality 5: Interlude II

2017-10-14T02:05:37.208Z · score: 12 (2 votes)

Instrumental Rationality 4.3: Breaking Habits and Conclusion

2017-10-12T23:11:18.127Z · score: 5 (4 votes)

Instrumental Rationality 4.2: Creating Habits

2017-10-12T02:25:06.007Z · score: 19 (8 votes)

The Recognizing vs Generating Distinction

2017-10-09T16:56:09.379Z · score: 18 (4 votes)

Instrumental Rationality 4.1: Modeling Habits

2017-10-09T01:21:41.396Z · score: 18 (8 votes)

Instrumental Rationality 3: Interlude I

2017-10-07T05:22:09.663Z · score: 18 (8 votes)

Instrumental Rationality 2: Planning 101

2017-10-06T14:23:06.190Z · score: 25 (11 votes)

Instrumental Rationality 1: Starting Advice

2017-10-05T04:37:21.557Z · score: 21 (14 votes)

The Best Self-Help Should Be Self-Defeating

2017-09-26T06:16:32.059Z · score: 14 (6 votes)

Instrumental Rationality Sequence Finished! (w/ caveats)

2017-09-09T01:49:53.109Z · score: 5 (5 votes)

Habits 101: Techniques and Research

2017-08-22T10:54:45.552Z · score: 5 (5 votes)

Bridging the Intention-Action Gap (aka Akrasia)

2017-08-01T22:31:31.577Z · score: 1 (1 votes)

Daniel Dewey on MIRI's Highly Reliable Agent Design Work

2017-07-09T04:35:44.356Z · score: 10 (10 votes)

The Use and Abuse of Witchdoctors for Life

2017-06-24T20:59:06.983Z · score: 4 (4 votes)

Instrumental Rationality 1: Starting Advice

2017-06-18T18:43:23.759Z · score: 13 (13 votes)

Recovering from Failure

2017-06-11T21:57:01.918Z · score: 0 (0 votes)

Probabilistic Programming and Bayesian Methods for Hackers

2017-05-22T21:15:25.129Z · score: 1 (2 votes)

Instrumental Rationality Sequence Update (Drive Link to Drafts)

2017-05-19T04:01:09.368Z · score: 2 (3 votes)

Decision Theories in Real Life

2017-05-13T01:47:41.790Z · score: 2 (3 votes)

There is No Akrasia

2017-04-30T15:33:39.710Z · score: 30 (22 votes)