Posts

lincolnquirk's Shortform 2023-02-22T15:05:43.390Z
Why I Think Abrupt AI Takeoff 2022-07-17T17:04:06.038Z
Unlock the Door 2021-10-29T23:45:43.273Z
Book Review: A Pattern Language by Christopher Alexander 2021-10-15T01:11:24.327Z
Process Orientation 2021-03-21T15:21:32.085Z
Enabling Children 2021-03-03T00:02:01.350Z
Could someone please start a bright home lighting company? 2019-11-26T19:20:04.622Z
[link] "Paper strips" memory technique 2018-10-06T12:11:35.248Z
Moral frameworks and the Harris/Klein debate 2018-05-03T09:49:03.626Z
[LINK] How to increase conscientiousness 2013-10-04T15:26:16.002Z
[LINK] Why Your Customers Would Be Happier If You Charged More 2012-09-21T20:10:39.525Z
Heading off a near-term AGI arms race 2012-08-22T14:23:58.382Z
How to enjoy being wrong 2011-07-27T05:48:01.005Z

Comments

Comment by lincolnquirk on The Best Tacit Knowledge Videos on Every Subject · 2024-04-05T20:34:16.409Z · LW · GW

For home cooking I would like to recommend J. Kenji Lopez-Alt (https://www.youtube.com/@JKenjiLopezAlt/videos). He's a well-loved professional chef who writes science-y cooking books, and his youtube channel is a joy because it's mostly just low production values: him in his home kitchen, making delicious food from simple ingredients, just a few cuts to speed things up.

Comment by lincolnquirk on Sharing Information About Nonlinear · 2023-09-12T03:41:49.732Z · LW · GW

I'm sorry you feel that way. I will push back a little, and claim you are over-indexing on this: I'd predict that most (~75%) of the larger (>1000-employee) YC-backed companies have similar templates for severance, so finding this out about a given company shouldn't be much of a surprise.

I did a bit of research to check my intuitions + it does seem like non-disparagement is at least widely advised (for severance specifically and not general employment), e.g., found two separate posts on the YC internal forums regarding non-disparagement within severance agreements:

"For the major silicon valley law firms (Cooley, Fenwick, OMM, etc) non disparagement is not in the confidentiality and invention assignment agreement [employment agreement], and usually is in the separation and release [severance] template."

(^ this person also noted that it would be a red flag to find non-disparagement in the employment agreement.)

"One thing I’ve learned - even when someone has been terminated with cause, a separation agreement [which includes non-disparagement] w a severance can go a long way."

Comment by lincolnquirk on Sharing Information About Nonlinear · 2023-09-11T23:24:06.447Z · LW · GW

Jeff is talking about Wave. We use a standard form of non-disclosure and non-disparagement clauses in our severance agreements: when we fire or lay someone off, getting severance money is gated on not saying bad things about the company. We tend to be fairly generous with our severance, so people in this situation usually prefer to sign and agree. I think this has successfully prevented (unfair) bad things from being said about us in a few cases, but I am reading this thread and it does make me think about whether some changes should be made.

I also would re-emphasize something Jeff said - that these things are quite common - if you just google for severance package standard terms, you'll find non-disparagement clauses in them. As far as I am aware, we don't ask current employees or employees who are quitting without severance to not talk about their experience at Wave.

Comment by lincolnquirk on The Economics of the Asteroid Deflection Problem (Dominant Assurance Contracts) · 2023-08-30T14:05:30.037Z · LW · GW

In my view you have two plausible routes to overcoming the product problem, neither of which is solved (primarily) by writing code.

Route A would be social proof: find a trusted influencer who wants to do a project with DACs. Start by brainstorming various types of projects that would most benefit from DACs, aiming to find an idea which an (ideally) narrow group of people would be really excited about, that demonstrates the value of such contracts, led by a person with a lot of 'star power'. Most likely this would be someone who would be likely to raise quite a lot of money through a traditional donation/kickstarter-type drive, but instead they decide to demo the DAC (and in doing so make a good case for it).

Route B is to focus on comms. Iterate on the message. Start by explaining it to non-economist friends, then graduate to focus groups. It's crucial to try to figure out how to most simply explain the idea in a sentence or two, such that people understand and don't get confused by it.

I'm guessing you'll need to follow both these routes, but you can follow them simultaneously and hopefully learn cross-useful things while doing so.

Comment by lincolnquirk on The Economics of the Asteroid Deflection Problem (Dominant Assurance Contracts) · 2023-08-29T22:05:41.749Z · LW · GW

I like the idea of getting more people to contribute to such contracts. Not thrilled about the execution. I think there is a massive product problem with the idea -- people don't understand it, think it is a scam, etc. If your efforts were more directed at the problem of getting people to understand and be excited about crowdfunding contracts like this, I would be a lot more excited.

Comment by lincolnquirk on Why was the AI Alignment community so unprepared for this moment? · 2023-07-19T00:23:18.409Z · LW · GW

Mild disagree: I do think x-risk is a major concern, but seems like people around DC tend to put 0.5-10% probability mass on extinction rather than the 30%+ that I see around LW. This lower probability causes them to put a lot more weight on actions that have good outcomes in the non extinction case. The EY+LW frame has a lot more stated+implied assumptions about uselessness of various types of actions because of such high probability on extinction.

Comment by lincolnquirk on Why was the AI Alignment community so unprepared for this moment? · 2023-07-18T22:29:05.668Z · LW · GW

Your question is coming from within a frame (I'll call it the "EY+LW frame") that I believe most of the DC people do not heavily share, so it is kind of hard to answer directly. But yes, to attempt an answer, I've seen quite a lot of interest (and direct policy successes) in reducing AI chips' availability and production in China (eg via both CHIPS act and export controls), which is a prerequisite for US to exert more regulatory oversight of AI production and usage. I think the DC folks seem fairly well positioned to give useful inputs into further AI regulation as well.

Comment by lincolnquirk on Why was the AI Alignment community so unprepared for this moment? · 2023-07-18T20:11:29.215Z · LW · GW

I've been in DC for ~ the last 1.5y and I would say that DC AI policy has a good amount of momentum, I doubt it's particularly visible on twitter but also it doesn't seem like there are any hidden/secret missions or powerful coordination groups (if there are, I don't know about it yet). I know ~10-20 people decently well here who work on AI policy full time or their work is motivated primarily by wanting better AI policy, and maybe ~100 who I have met once or twice but don't see regularly or often; most such folks have been working on this stuff since before 2022; they all have fairly normal-seeming thinktank- or government-type jobs.

They don't mostly spend time on LW (although certainly a few of them do). Many do spend time on Twitter, and they do read lots of AI related takes from LW-influenced folks. They have meetup groups related to AI policy. I guess it looks pretty much as I was expecting before I came here. Happy to answer further questions that don't identify specific people, just because I don't know how many of them want to be pointed-at on LW.

Comment by lincolnquirk on Change my mind: Veganism entails trade-offs, and health is one of the axes · 2023-06-05T14:14:21.286Z · LW · GW

Not who you're responding to, but I've just written up my vegan nutrition tips and tricks: http://www.lincolnquirk.com/2023/06/02/vegan_nutrition.html

Comment by lincolnquirk on Change my mind: Veganism entails trade-offs, and health is one of the axes · 2023-06-02T15:10:47.255Z · LW · GW

If you have energy for this, I think it would be insanely helpful!

Comment by lincolnquirk on Change my mind: Veganism entails trade-offs, and health is one of the axes · 2023-06-02T15:10:12.106Z · LW · GW

Thanks for writing this. I think it's all correct and appropriately nuanced, and as always I like your writing style. (To me this shouldn't be hard to talk about, although I guess I'm a fairly recent vegan convert and haven't been sucked into whatever bubble you're responding to!)

Comment by lincolnquirk on Lessons learned from offering in-office nutritional testing · 2023-05-16T19:49:54.723Z · LW · GW

Thanks for doing this! These results may affect my supplementation strategy.

My recent blood tests (unrelated to this blog post) -- if you have any thoughts on them let me know, I'd be curious what your threshold for low-but-not-clinical is.

  • Hemoglobin - 14.8 g/dL
  • Vitamin D, 25-Hydroxy - 32.7 ng/mL
  • Vitamin B12 - 537 pg/mL

(I have other results I can send you privately if you want, from comp metabolic panel + cbc + lipid panel + D + B12; but didn't think to ask for iron. Is it worth going back to ask for this? or might iron be under a name I don't recognize?)

I'm vegan and have been solidly for > 1 year. Generally feel good, no particular fatigue except sleepiness after I eat carbs for lunch. I supplement B12, omega-3 EPA+DHA algae oil, creatine and occasional D3 gummies.

Comment by lincolnquirk on lincolnquirk's Shortform · 2023-02-22T15:05:43.699Z · LW · GW

Tim Urban's new book, What's Our Problem, is out as of yesterday. I've started reading it and it's good so far, and very applicable to rationality training. waitbutwhy.com

Comment by lincolnquirk on LW Beta Feature: Side-Comments · 2022-11-24T13:58:52.772Z · LW · GW

Excited about this!

Points of feedback:

  1. I don't like to have to scroll my screen horizontally to read the comment. (I notice there's a lot of perfectly good unused white space on the left side; comments would probably fit horizontally if you pushed everything to the left!)
  2. Sometimes when you mouse over the side-comment icon, it tries to scroll the page to make the comment readable. This is very surprising and makes me lose my place.
  3. Hovering over the icon makes the comment appear briefly. If I then want to scroll in order to read the comment, there seems to be no way to 'stay hovered' -- I have to click and toggle it, to make the comment stick around so I can actually read it. (This plus being forced to scroll the screen makes the hover feature kind of useless.)

Overall, feeling optimistic though, and will probably use this.

Comment by lincolnquirk on Don't expect AGI anytime soon · 2022-10-11T13:15:35.364Z · LW · GW

I think your argument is wrong, but interestingly so. I think DL is probably doing symbolic reasoning of a sort, and it sounds like you think it is not (because it makes errors?)

Do you think humans do symbolic reasoning? If so, why do humans make errors? Why do you think a DL system won't be able to eventually correct its errors in the same way humans do?

My hypothesis is that DL systems are doing a sort of fuzzy finite-depth symbolic reasoning -- it has capacity to understand the productions at a surface level and can apply them (subject to contextual clues, in an error-prone way) step by step, but once you ask for sufficient depth it will get confused and fail. Unlike humans, feedforward neural nets can't think for longer and churn step by step yet; but if someone were to figure out a way to build a looping option into the architecture then I won't be surprised to see DL systems which can go a lot further on symbolic reasoning than they currently do.

Comment by lincolnquirk on Will you let your kid play football? · 2022-10-04T19:31:44.695Z · LW · GW

What is Pop Warner in this context? I have googled it and it sounds like he was one of the founders of modern American football, but I don't understand what it is in contrast to. Is there some other (presumably safer) ruleset?

Comment by lincolnquirk on My emotional reaction to the current funding situation · 2022-09-10T01:07:32.981Z · LW · GW

(Inside-of-door-posted hotel room prices are called "rack rates" and nobody actually pays those. This is definitely a miscommunication.)

Comment by lincolnquirk on What does moral progress consist of? · 2022-08-19T00:39:30.172Z · LW · GW

I am guilty of being a zero-to-one, rather than one-to-many, type person. It seems far easier and more interesting to me, to create new forms of progress of any sort, rather than convincing people to adopt better ideas.

I guess the project of convincing people seems hard? Like, if I come up with something awesome that's new, it seems easier to get it into people's hands, rather than taking an existing thing which people have already rejected and telling them "hey this is actually cool, let's look again".

All that said, I do find this idea-space intriguing partly thanks to this post - it makes me want to think of ways of doing more one-to-many type stuff. I've been recently drawn into living in DC and I think the DC effective altruism folks are much more on the one-to-many side of the world.

Comment by lincolnquirk on Proposal: Consider not using distance-direction-dimension words in abstract discussions · 2022-08-09T21:00:49.024Z · LW · GW

Upvoted for raising something to conscious attention, that I have never previously considered might be worth paying attention to.

(Slightly grumpy that I'm now going to have a new form of cognitive overhead probably 10+ times per day... these are the risks we take reading LW :P)

Comment by lincolnquirk on How Promising is Theoretical Research on Rationality? Seeking Career Advice · 2022-07-26T02:33:46.478Z · LW · GW

Look, I don’t know you at all. So please do ignore me if what I’m saying doesn’t seem right, or just if you want to, or whatever.

I’m a bit worried that you’re seeking approval, not advice? If this is so, know that I for one approve of your chosen path. You are allowed to spend a few years focusing on things that you are passionate about, which (if it works out) may result in you being happy and productive and possibly making the world better.

If you are in fact seeking advice, you should explain what your goal is. If your goal is to make the maximum impact possible — it’s worth at least hundreds of hours trying to see if you can learn more & motivate yourself along a path which seems like it combines high impact with personal resonance. I wouldn’t discount philosophy along this angle, but (for example) it sounds like you may not know that much about the potential of policy careers; there are plenty that do not require particularly strong mathematical skills (… or even any particularly difficult skills beyond some basic extraversion, resistance to boredom and willingness to spend literal decades grinding away within bureaucracies).

If your goal is to be happy, I think you will be happy doing philosophy, and I think you have a potential to make a huge impact that way. Certainly there are a decent number of full-time philosophers within effective altruism who I have huge respect for (Macaskill, Ord, Bostrom, Greaves, and Trammell jump to mind). Plus, you can save a few hundred hours, which seems pretty important if you might already know the outcome of your experimentation!

Comment by lincolnquirk on Why I Think Abrupt AI Takeoff · 2022-07-18T02:54:49.969Z · LW · GW

Thanks! This is very helpful, and yes, I did mean to refer to grokking! Will update the post.

Comment by lincolnquirk on What Are You Tracking In Your Head? · 2022-06-28T21:18:01.992Z · LW · GW

Nice post!

One of my fears is that the True List is super long, because most things-being-tracked are products of expertise in a particular field and there are just so many different fields.

Nevertheless:

  • In product/ux design, tracking the way things will seem to a naive user who has never seen the product before.
  • In navigation, tracking which way north is.
  • I have a ton of "tracking" habits when writing code:
    • types of variables (and simulated-in-my-head values for such)
    • refactors that want to be done but don't quite have enough impetus for yet
    • loose ends, such as allocated-but-not-freed resources, or false symmetry (something that looks like it should be symmetric but isn't in some critical way), or other potentially-misleading things that need to be explained
    • [there are probably a lot more of these that I am not going to write down now]
Comment by lincolnquirk on Let's See You Write That Corrigibility Tag · 2022-06-24T18:30:08.147Z · LW · GW

Here's my attempt. I haven't read any of the other comments or the tag yet. I probably spent ~60-90m total on this, spread across a few days.

On kill switches

  • low impact somehow but I don’t know how
  • Go slow enough so that people can see what you’re doing
  • Have a bunch of "safewords" and other kill-switches installed at different places, some hopefully hard-to-reach by the AI. Test them regularly, and consider it a deadly flaw if one stops working.

On the AI accurately knowing what it is doing, and pointing at things in the real world

  • watch all the metrics (!)
  • Predict all the metrics you watch, and ask humans about any anomalous metrics that you are watching
  • group inputs and outputs separately and treat inputs as sacred. Perhaps by having an epistemic module which is incentivized by producing true predictions about the world. Make the epistemic module the one that gets all the juice. Planning module should not have any way to influence it.

On responding predictably

  • Require inputs of some kind to get more outputs (something about control theory?) - like a power steering wheel, or an ebike, it helps the human by amplifying the motions you make but not actually Doing Anything on its own
  • Have metrics look smooth on an input/output response curve. No sharp edges. let the humans be the one to turn the knobs.
  • Try things. Try things on a small scale before going big. (define "big"? maybe look at the metrics!)

On epistemology and ontology

  • Ask for feedback both about things you tried; and experiments you want to run
  • make the epistemic module have a stable ontology. If it wants to change the ontology, it needs to create a new epistemic module, and leave the old one running. The new module has to be able not only to predict physical things about the world but also what the old model would say about things being pointed-to and tagged as important in the old models ontology.
  • have a concept of goal metrics and side effects. Try to predict what the human will say about which metrics are goals and which are side effects. (hmm..this is incoherent because you can come up with many irrelevant metrics… what even is a relevant metric. Maybe the system should be grouping metrics using some kind of natural unsupervised clustering system and then trying to predict the prioritization of metric groups? And then to explain the groups it gives samples of the actual thing it’s measuring. Still difficult to make sense of this. )

On planning

  • Bound or penalize plans in complexity and time.
    • "complexity" is hard to define here, and isn't Kolmogorov I don't think.
  • Harshly penalize plans which lean on predicting what other agents will do. Ban plans which depend on prediction of what your operator will do. (AI can still predict what the operator will think of the plan, but that prediction should not feed into any plan itself.)
  • Separate planning from other parts of AI cognition, such that you are never confused about whether you are making a plan or doing other things, such as predicting things about the world, or executing a plan.
  • Don't execute plans at all
  • Train the AI to identify the purpose(s) of each step in the plan. Give negative feedback when steps lead to outcomes that are not identified ahead of time.

On transparency

  • A part of the AI should be trained to explicate what the rest of the AI is "doing"/focused on at all times. Find a way to make this into a feedback cycle.
  • AI should be arranged so it is mostly only doing one thing at a time, and that doesn't switch its focus too many times each minute.
Comment by lincolnquirk on On A List of Lethalities · 2022-06-13T18:13:42.495Z · LW · GW

I notice that I am extremely surprised by your internship training. Its existence, its lessons and the impact it had on you (not you specifically, just a person who didn't come in with that mindset) are all things I don't think I would have predicted. I would be thrilled if you would write as much as you can bring yourself to about this, braindump format is fine, into a top level post!

Comment by lincolnquirk on Leaving Google, Joining the Nucleic Acid Observatory · 2022-06-10T22:58:41.891Z · LW · GW

Congrats, I'm excited about this!

Comment by lincolnquirk on Consequentialist veganism · 2022-03-28T03:35:19.562Z · LW · GW

I've been turning this over in my head for a while now. (Currently eating mostly vegan fwiw, but I am not sure if this is the right decision.)

I think the main argument against veganism is that it actually incurs quite a large cost. Being vegan is a massive lifestyle change with ripple effects that extend into one's social life. This argument falls under your "there are higher-impact uses of your (time/energy/money/etc.)", but what you wrote doesn't capture the reasons why this is important.

most of us do not have good reason to treat this as a zero-sum game in which each attempt to do good in the world must crowd out another. For one thing, we're nowhere near putting all available resources into our efforts to do good, so we can simply choose to expand that budget.

I am reminded of Zvi's Slack post (ctrl+f "afford"). Attention is a very scarce resource. If I am spending all my attention on important things, I cannot also spend attention on creating a whole new diet, finding friends who won't mock me, learning how to cook all new things, etc. On the other hand, animal welfare offsets are probably quite cheap.

For another, our psychology is complicated, and making a moral effort can just as easily increase our capacity to make further such efforts as deplete it.

Indeed, and this is actually why I have become mostly vegan in recent months; but it is not going to be true for everyone. My current decision to eat mostly vegan except when inconvenient feels somehow indulgent.

I wrote more about how I am trying to be vegan: http://www.lincolnquirk.com/2022/02/15/vegan.html

Comment by lincolnquirk on Where can I read about the Roko's Basilisk drama? · 2022-01-30T16:50:24.420Z · LW · GW

This community has a virtue of taking weird ideas seriously. Roko came up with a weird idea which, the more seriously you took it, the more horrifying it became. This was deemed an info hazard, and censored in some way, I don't know how. But the people who didn't take it seriously in the first place weren't horrified by the idea and thus were confused about why it should have been censored, and thus boosted the Streisand effect.

Comment by lincolnquirk on Mildly Photochromic Lenses? · 2022-01-29T18:09:15.159Z · LW · GW

I had photochromics for several years. I found them mildly-helpful-and-mostly-unobjectionable in the summer, but ridiculously annoying in the winter (when they both tend to be darker because of low-altitude sun, and the temperature makes them clear up slower once you move inside).

Also, I was relentlessly mocked by the fashion police. :P

Ultimately I moved away from them.

Comment by lincolnquirk on Consume fiction wisely · 2022-01-21T21:06:32.463Z · LW · GW

I downvoted this. I usually like the concise writing style exhibited in this essay (similar to lsusr, paul graham, both of whom I like) , but I apparently only like it when I think it's correct. :P

I especially downvoted because I think it is fairly likely to attract low-quality discussion. A differently-written version of a similar but perhaps more nuanced point, with better fleshed-out examples of why given works are net helpful or net harmful, would be a better post. I am sympathetic to the general idea of the post!

Comment by lincolnquirk on Why did computer science get so galaxy-brained? · 2021-12-27T18:16:03.053Z · LW · GW

I think there's something about programming that attracts the right sort of people. What could that be? Well, programming has very tight feedback loops, which make it fun. You can "do a lot": one's ability to gain power over the universe, if you will, is quite high with programming. I'd guess a combination of these two factors.

Comment by lincolnquirk on A good rational fiction about IT-inspired magic system? · 2021-12-27T16:32:10.772Z · LW · GW

The Wizard's Bane series by Rick Cook. The basic idea is great: a Silicon Valley programmer is transported into a magical universe where he has to figure out how to apply programming to a magic system. Caveat lector: the writing is not the best quality, it's a bit juvenile, but still a light, enjoyable read :)

Comment by lincolnquirk on Experiences raising children in shared housing · 2021-12-21T18:06:43.643Z · LW · GW

This is great! Thanks for sharing!

Comment by lincolnquirk on Ngo's view on alignment difficulty · 2021-12-16T00:11:22.652Z · LW · GW

A fair question. I don't think it is established, exactly, but the plausible window is quite narrow. For example, if nanomachinery were easy, we would already have that technology, no? And we seem quite near to AGI.

Comment by lincolnquirk on Soares, Tallinn, and Yudkowsky discuss AGI cognition · 2021-11-29T23:07:58.592Z · LW · GW

evolution would love superintelligences whose utility function simply counts their instantiations! so of course evolution did not lack the motivation to keep going down the slide. it just got stuck there (for at least ten thousand human generations, possibly and counterfactually for much-much longer). moreover, non evolutionary AI’s also getting stuck on the slide (for years if not decades; median group folks would argue centuries) provides independent evidence that the slide is not too steep (though, like i said, there are many confounders in this model and little to no guarantees).

Evolution got stuck on the slide with humans because cultural evolution outcompeted biological evolution, because of cultural evolution's ability to make immediate direct impacts on small tribes in hunter-gatherer environment within a few short generations (from the first chapter of Secret Of Our Success) and the high-order bit in biological evolution suddenly became "how efficient is cultural evolution".

(Non evolutionary AIs don't seem stuck on the slide at all to me.)

Comment by lincolnquirk on Sasha Chapin on bad social norms in rationality/EA · 2021-11-18T10:18:01.409Z · LW · GW

I don’t think I agree that this is made-up though. You’re right that the quotes are things people wouldn’t say but they do imply it through social behavior.

I suppose you’re right that it’s hard to point to specific examples of this happening but that doesn’t mean it isn’t happening, just that it’s hard to point to examples. I personally have felt multiple instances of needing to do the exact things that Sasha writes about - talk about/justify various things I’m doing as “potentially high impact”; justify my food choices or donation choices or career choices as being self-improvement initiatives; etc.

this article points at something real

Comment by lincolnquirk on Ngo and Yudkowsky on alignment difficulty · 2021-11-16T11:05:02.256Z · LW · GW

I'd like to express my gratitude and excitement (and not just to you, Rob, though your work is included in this):

Deep thanks to everyone involved for having the discussion, writing up and formatting, and posting it on LW. I think this is some of the more interesting and potentially impactful stuff I've seen relating to AI alignment in a long while.

(My only thought is... why hasn't a discussion like this occurred sooner? Or has it, and it just hasn't made it to LW?)

Comment by lincolnquirk on Dating profiles from first principles: heterosexual male profile design · 2021-10-25T13:59:29.591Z · LW · GW

Regardless of the precise mechanism, Tinder almost certainly shows more attractive people more often. If it didn't, it would have a retention problem because there are lots of people who swipe tinder to fantasize about matching with hot people, and they wouldn't get enough hot people to keep them going. Most likely, Tinder has determined a precise ratio of "hot people" and "people in your league" to show you, in order to keep you swiping.

Given the existence of the incentive and likelihood that Tinder et al. would follow such an incentive, it makes sense to try to have your profile be more generally attractive so you get shown to more people.

Comment by lincolnquirk on Book Review: A Pattern Language by Christopher Alexander · 2021-10-15T01:20:51.040Z · LW · GW

Use the table of contents / "summary of the language" section.

For your project I would recommend skipping to 28 and then going from there, and skipping patterns which don't seem relevant.

Comment by lincolnquirk on How to think about and deal with OpenAI · 2021-10-13T23:33:36.052Z · LW · GW

Yes: A far higher % of OpenAI reads this forum than the other orgs you mentioned. In some sense OpenAI is friends with LW, in a way that is not true for the others.

Comment by lincolnquirk on How to think about and deal with OpenAI · 2021-10-13T23:32:07.691Z · LW · GW

What should be done instead of a public forum? I don't necessarily think there needs to be a "conspiracy", but I do think that it's a heck of a lot better to have one-on-one meetings with people to convince them of things. At my company, when sensitive things need to be decided or acted on, a bunch of slack DMs fly around until one person is clearly the owner of the problem; they end up in charge of having the necessary private conversations (and keeping stakeholders in the loop). Could this work with LW and OpenAI? I'm not sure.

Comment by lincolnquirk on How to think about and deal with OpenAI · 2021-10-13T23:05:35.992Z · LW · GW

Ineffective, because the people arguing on the forum are lacking knowledge about the situation. They don't understand OpenAI's incentive structure, plan, etc. Thus any plans they put forward will be in all likelihood useless to OpenAI.

Risky, because (some combination of):

  • it is emotionally difficult to hear that one of your friends is plotting against you (and openAI is made up of humans, many of whom came out of this community)
    • it's especially hard if your friend is misinformed and plotting against you; and I think it likely that the openAI people believe that Yudkowsky/LW commentators are misinformed or at least under-informed (and they are probably right about this)
  • to manage that emotional situation, you may want to declare war back on them, cut off contact, etc.; any of these actions if declared as an internal policy would be damaging to the future relationship between openAI and the LW world
  • openAI has already had a ton of PR issues over the last few years and so they probably have a pretty well developed muscle for dealing internally with bad PR, which this would fall under. If true, the muscle probably looks like internal announcements with messages like "ignore those people/stop listening to them, they don't understand what we do, we're managing all these concerns and those people are over indexing on them anyway"
  • the evaporative cooling effect may eject some people who were already on the fence about leaving, but the people who remain will be more committed to the original mission, more "anti LW" and less inclined to listen to us in the future
  • hearing bad arguments makes one more resistant to similar (but better) arguments in the future

I want to state for the record that I think OpenAI is sincerely trying to make the world a better place, and I appreciate their efforts. I don't have a settled opinion on the sign of their impact so far.

Comment by lincolnquirk on How to think about and deal with OpenAI · 2021-10-10T14:09:58.782Z · LW · GW

I'd like to put in my vote for "this should not be discussed in public forums". Whatever is happening, the public forum debate will have no impact on it; but it does create the circumstances for a culture war that seems quite bad.

Comment by lincolnquirk on Common knowledge about Leverage Research 1.0 · 2021-10-05T15:57:32.286Z · LW · GW

When I learned it from Geoff in 2011, they were recommending yEd Graph Editor. The process is to generally write things you do or want to do as nodes, and then connect them to each other using "achieves or helps to achieve" edges (i.e., if you go to work, that achieves making money, which achieves other things you want).

Comment by lincolnquirk on Common knowledge about Leverage Research 1.0 · 2021-10-05T15:56:17.196Z · LW · GW

I believe this. Aversion factoring is a separate insight from goal factoring.

Comment by lincolnquirk on What should one's policy regarding dental xrays be? · 2021-09-18T08:30:49.062Z · LW · GW

XKCD says that the dental X-ray (5 μSv) is half the average daily background radiation dose (10 μSv), and 1/8th of a cross country flight (40 μSv). To me this means that the radiation exposure is quite irrelevant in the grand scheme of things. (https://xkcd.com/radiation/)

If this were false, it would presumably be because dental X-rays are especially harmful in some way that isn't just "because of radiation".

Comment by lincolnquirk on Outline of Galef's "Scout Mindset" · 2021-08-11T14:47:05.582Z · LW · GW

I didn't read Scout Mindset yet, but I've listened to Julia's interviews on podcasts about it, and I have read the other books that Rob mentions in that paragraph.

The reason I nodded when Rob wrote that was that Julia's memetics are better. Her ideas are written in a way which stick in one's mind, and thus spread more easily. I don't think any of those other sources are bad -- in fact I get more from them than I expect to from Scout Mindset -- but Scout Mindset is more practically oriented (and optimized for today's political climate) in a way which those other books are not.

It also operates at a different, earlier level in the "EA Funnel": the level at which you can make people realize that more is possible. Those other books already require someone to be asking "how can I Do Good Better?" before they'll pick it up.

Comment by lincolnquirk on Two non-obvious lessons from microcovid.org · 2021-06-24T10:55:10.373Z · LW · GW

Thanks for writing this, and for writing the software! Microcovid was quite impactful in my own life and the software is shockingly thoughtful — you put so many tiny little details in to help me decide (oh, it’s a taxi ride and I don’t think the driver is going to wear a mask, but I can open the window…)

I think this must be a result of your & your team’s hard work talking to nonrationalists, as you note, but I also think you must have really good product instincts. Just talking to people is not, in my view, enough to produce a product that thoughtful — you also have to figure out what to do with the data. Nice work :)

Comment by lincolnquirk on Raemon's Shortform · 2021-06-17T16:45:46.370Z · LW · GW

I'm not Ray, but I'll take a stab --

The founder has a complete vision for the community/meetup/company/etc. They were able to design a thing that (as long as they continue putting in energy) is engaging, and they instinctively know how to change it so that it continues being great for participants.

The first successor has an incomplete, operational/keep-things-running-the-way-they-were type vision. They cargo-cult whatever the founder was doing. They don't have enough vision to understand the 'why' behind all the decisions. But putting your finger on their precise blind spot is quite hard. It's their "fault" (to the extent that we can blame anyone) that things go off the rails, but their bad decision-making doesn't actually have short term impacts that anyone can see. Instead, the impacts come all at once, once they disappear, and there becomes common knowledge that it was a house of cards the whole time.

(or something. my models are fairly imprecise on this.)

Anyway, why did the founder get fooled into anointing the first successor even though they don't have the skills to continue the thing? My guess is that there's a fairly strong selection effect for founders combined with "market fit" -- founders who fail to reach this resonant frequency don't pick successors, they just fail. Whatever made them great at building this particular community doesn't translate into skills at picking a successor, and that resonance may not happen to exist in any other person. Another founder-quality person would not necessarily have resonated with the existing community's frequency, so there could also be an anti-selection effect there.

Comment by lincolnquirk on What are your greatest one-shot life improvements? · 2021-06-05T23:57:01.527Z · LW · GW

I didn’t do it any more. I forgot about it next time I showered.

Comment by lincolnquirk on Peekskill Lyme Incidence · 2021-06-02T12:55:38.583Z · LW · GW

Sorry I don’t have good answers:

  1. I don’t think so, typically you’re “cured” if you get antibiotics early on and don’t see any more symptoms
  2. diagnosis was only from the characteristic “ring” rash around bite site. The nurse said that it was the “best” (clearest example) of one she had ever seen
  3. Not concerned, given i caught it quickly and haven’t seen any other symptoms