Posts

Panel discussion on AI consciousness with Rob Long and Jeff Sebo 2023-09-09T03:38:17.662Z
Podcast (+transcript): Nathan Barnard on how US financial regulation can inform AI governance 2023-08-08T21:46:05.299Z
Announcing the EA Archive 2023-07-06T13:49:17.387Z
Most Ivy-smart students aren't at Ivy-tier schools 2022-08-07T03:18:02.298Z
Half-baked alignment idea: training to generalize 2022-06-19T20:16:43.735Z
Aaron Bergman's Shortform 2022-06-15T07:11:15.053Z
For God's sake, Google it. 2021-08-10T23:52:24.408Z
Delta variant: we should probably be re-masking 2021-07-23T21:13:15.455Z
The topic is not the content 2021-07-06T00:14:26.106Z
Book review: Subtract: The Untapped Science of Less 2021-05-23T20:21:21.685Z
The case for hypocrisy 2021-05-13T03:36:16.863Z
Dumb dichotomies in ethics, part 2: instrumental vs. intrinsic values 2021-05-07T02:31:50.591Z
On silence 2021-04-30T03:15:23.009Z
What books are for: a response to "Why books don't work." 2021-04-21T23:18:09.393Z
The secret of Wikipedia's success 2021-04-14T22:18:52.871Z
Agreeing to Disagree: towards a better understanding of agreeableness 2021-03-29T03:14:23.966Z
Some things I've learned in college 2021-03-25T21:30:05.899Z
Direct effects matter! 2021-03-14T04:33:11.493Z
The Case for Logical Fallacies 2021-03-08T03:04:47.264Z
Societal Self-control 2021-02-20T03:05:21.271Z
Metric selection bias: why Moore's law is less important than you think 2021-02-08T00:21:20.494Z
Dumb Dichotomies in Ethics, Part 1: Intentions vs Consequences 2021-01-29T03:49:10.323Z
Efficiency Wages: A Double-Edged Sword 2021-01-11T04:43:09.201Z
Don't Use Your Favorite Weapon 2021-01-03T04:01:39.766Z

Comments

Comment by Aaron Bergman (aaronb50) on Announcing the EA Archive · 2023-07-09T15:00:33.655Z · LW · GW

I mean the reason is that I've never heard of that haha. Perhaps it should be

Comment by Aaron Bergman (aaronb50) on (My understanding of) What Everyone in Technical Alignment is Doing and Why · 2022-09-13T04:56:04.913Z · LW · GW

Ngl I did not fully understand this, but to be clear I don't think understanding alignment through the lense of agency is "excessively abstract." In fact I think I'd agree with the implicit default view that it's largely the single most productive lense to look through. My objection to the status quo is that it seems like the scale/ontology/lense/whatever I was describing is getting 0% of the research attention whereas perhaps it should be getting 10 or 20%.  

Not sure this analogy works, but if NIH was spending $10B on cancer research, I would (prima facie, as a layperson) want >$0 but probably <$2B spent on looking at cancer as an atomic-scale phenomenon, and maybe some amount at an even lower-scale scale

Comment by Aaron Bergman (aaronb50) on (My understanding of) What Everyone in Technical Alignment is Doing and Why · 2022-09-13T02:21:28.432Z · LW · GW

Note: I'm probably well below median commenter in terms of technical CS/ML understanding.  Anyway...

I feel like a missing chunk of research could be described as “seeing DL systems as ‘normal,’ physical things and processes that involve electrons running around inside little bits of (very complex) metal pieces” instead of mega-abstracted “agents.”

The main reason this might be fruitful is that, at least intuitively and to my understanding, failures like “the AI stops just playing chess really well and starts taking over the world to learn how to play chess even better” involve a qualitative change beyond just “the quadrillion parameters adjust a bit to minimize loss even more” that eventually cashes out in some very different way that literal bits of metal and electrons are arranged. 

And plausibly abstracting away from the chips and electrons means ignoring the mechanism that permits this change. Of course, this probably only makes sense if something resembling deep learning scales to AGI, but it seems that some very smart people think that it may! 

Comment by Aaron Bergman (aaronb50) on Most Ivy-smart students aren't at Ivy-tier schools · 2022-08-07T16:18:08.021Z · LW · GW

Banneker Key! Yeah I was in a very similar position, but basically made the opposite choice (largely because financial costs not internalized)

Comment by Aaron Bergman (aaronb50) on Most Ivy-smart students aren't at Ivy-tier schools · 2022-08-07T16:16:22.779Z · LW · GW

Yeah that's gotta be it, nice catch! 

Comment by Aaron Bergman (aaronb50) on What Are You Tracking In Your Head? · 2022-07-02T21:18:06.728Z · LW · GW

One answer to the question for me:

While writing, something close to "how does this 'sound' in my head naturally, when read, in an aesthetic sense?"

I've thought for a while that "writing quality" largely boils down to whether the writer has an intuitively salient and accurate intuition about how the words they're writing come across when read. 

Comment by Aaron Bergman (aaronb50) on Half-baked AI Safety ideas thread · 2022-06-24T21:43:23.418Z · LW · GW

Ah late to the party! This was a top-level post aptly titled "Half-baked alignment idea: training to generalize" that didn't get a ton of attention. 

Thanks to Peter Barnett and Justis Mills for feedback on a draft of this post. It was inspired by Eliezer's Lethalities post and Zvi's response.

Central idea: can we train AI to generalize out of distribution

I'm thinking, for example, of an algorithm like the following:

  1. Train a GPT-like ML system to predict the next word given a string of text only using, say, grade school-level writing (this being one instance of the object level)
    1. Assign the system a meta-level award based on how well it performs (without any additional training) at generalizing; in this case, that is, predicting the next word from more advanced, complex writing (perhaps using many independent tests of this task without updating/learning between each test, and allowing parameters to update only after the meta-level aggregate score is provided)
      • Note: the easy→hard generalization is not a necessary feature. Generalization could be from fiction→nonfiction writing or internet→native print text, for instance.
    2. After all these independent samples are taken, provide the AI its aggregate or average score as feedback
  2. (Maybe?) repeat all of step I on a whole new set of training and testing texts (e.g., using text from a different natural language like Mandarin)
    1. Repeat this step an arbitrary number of times
      • For example, using French text, then Korean, then Arabic, etc. 
  3. Each time a “how well did you generalize” score is provided (which is given once per natural language in this example), the system should improve at the general task of generalizing from simple human writing to more complex human writing, (hopefully) to the point of being able to perform well at generalizing from simple Hindi (or whatever) text to advanced Hindi prediction even if it had never seen advanced Hindi text before.  
  4. ^Steps 1-3 constitute the second meta-level of training an AI to generalize, but we can easily treat this process as a single training instance (e.g., rating how well the AI generalizes to Hindi advanced text after having been trained on doing this in 30 other languages) and iterate over and over again. I think this would look like:
    1. Running the analogs of steps 1-4 on generalizing from 
      • (a) simple text to advanced text in many languages
      • (b) easy opponents to hard ones across many games, 
      • (c) photo generation of common or general objects ("car") to rare/complex/specific ones ("interior of a 2006 Honda Accord VP"), across many classes of object
    2. And (hopefully) the system would eventually be able to generalize from simple Python code training data to advanced coding tasks even though it had never seen any coding at all before this. 

And, of course, we can keep on adding piling layers on. 

A few notes

  • I think the following is one way of phrasing what I hope might happen with method: we are using RL to teach an ML system how to do ML in such a way that it sacrifices some in-distribution predictive power for the ability to use its “knowledge” more generally without doing anything that seems dumb to us. 
  • Of course, there are intrinsic limits to any system’s ability to generalize. The system in question can only generalize using knowledge X if X exists as information in the object-level training provided to it. 
    • This limits what we should expect of the system.
      • For example, I am almost certain that even an arbitrarily smart system will not be able to generate coherent Mandarin text from English training data, because the meaning of Mandarin characters doesn’t exist as “latent knowledge” in even a perfect understanding of English. 

Anyone here know Python?

My hands-on experience with ML extends to linear regression in R and not an inch more, so I'm probably not the best person to test this theory out. I've heard some LWers know a bit of Python, though.

If that's you, I'd be fascinated and thankful to see if you can implement this idea using whatever data and structure you think would work best, and would be happy to collaborate in whatever capacity I can. 

 

Appendix: a few brief comments (from someone with much more domain knowledge than me) and responses (from me):

Comment

Is this just the same as training it on this more complex task (but only doing one big update at the end, rather than doing lots of small updates)?

Response (which may help to clarify why I believe the idea might work)

I don't think so, because the parameters don't change/update/improve between each of those independent tests. Like GPT-3 in some sense has a "memory" of reading Romeo and Juliet, but that's only because its parameters updated as a result of seeing the text.

But also I think my conception depends on the system having "layers" of parameters corresponding to each layer of training. 

So train on simple English-->only "Simple English word generation" parameters are allowed to change...but then you tell it how well it did at generalizing out of distribution, and now only its "meta level 1 generalization" parameters are allowed to change.

Then you do the whole thing again but with German text, and its "Meta level 1 generalization" parameters are allowed to change again using SGD or whatever. If this works, it will be the reason why it can do well at advanced Hindi text without ever having read advanced Hindi.

Treat this whole process as the object level, and then it updates/improves "meta level 2 generalization" parameters.


Comment: 

This looks vaguely like curriculum learning, which apparently doesn't really work in LLMs https://arxiv.org/abs/2108.02170, I think a similar experiment would be like train on simple+advanced text for English, French, Mandarin etc, but only simple Hindi, and then see if it can do complex Hindi. 

Response 

I think that's a pretty different thing because there are no meta level parameters. Seems like fundamentally just a flavor of normal RL

Or do pretraining with English, French, Mandarin, and Hindi, but only do fine tuning with English, French, Mandarin, and see if it can then do the tasks it was fine tuned for in Hindi. 

My prediction: it learns to generalize a bit (the scores on the novel Hindi tasks are higher than if there was no fine tuning with the other languages) but worse than the other languages generalize. As the models are scaled up, this 'generalization gap' gets smaller.

Seems like this might depend on the relative scaling of different meta level parameters (which I described above)? 

Like for example whenever you scale the # of object level params by a factor of 2, you have to scale the number of nth meta level parameters by 2^(n+1).

Comment by Aaron Bergman (aaronb50) on Announcing the LessWrong Curated Podcast · 2022-06-23T07:51:41.989Z · LW · GW

Thank you, Solenoid! The SSC podcast is the only reason I to consume all of posts like Biological Anchors: A Trick That Might Or Might Not Work

Comment by Aaron Bergman (aaronb50) on Half-baked alignment idea: training to generalize · 2022-06-20T20:40:50.859Z · LW · GW

Thanks. It's similar in one sense, but (if I'm reading the paper right) a key difference is that in the MAML examples, the ordering of the meta-level and object level training is such that you still wind up optimizing hard for a particular goal. The idea here is that the two types of training function in opposition, as a control system of sorts, such that the meta-level training should make the model perform worse at the narrow type of task it was trained on. 

That said, for sure, the types of distribution shift thing is an issue. It seems like this meta-level bias might be less bad than at the object level, but I have no idea. 

Comment by Aaron Bergman (aaronb50) on Aaron Bergman's Shortform · 2022-06-15T07:11:15.976Z · LW · GW

Training to generalize (and training to train to generalize, etc.)

Inspired by Eliezer's Lethalities post and Zvi's response:

Has there been any research or writing on whether we can train AI to generalize out of distribution

I'm thinking, for example:

  1. Train a GPT-like ML system to predict the next word given a string of text only using, say, grade school-level writing (this is one instance of the object level
    1. Assign the system a meta-level award based on how well it performs (without any additional training) at generating the next word from more advanced, complex writing (likely using many independent tests of this task) 
    2. After all these independent samples are taken, provide the its aggregate or average score as feedback
  2. (Maybe?) repeat steps I and I.I on a whole new set of training and testing texts (e.g., using text from a different natural language like Mandarin), and do this step and arbitrary number of times
    1. Repeat this step using French text, then Korean, then Arabic, etc. 
  3.  After each of the above steps, the system should (I assume) improve at the general task of generalizing from simple human writing to more complex human writing, (hopefully) to the point of being able to perform well at generalizing from simple Hindi (or whatever) text to advanced Hindi prediction even if it had never seen advanced Hindi text before.  
  4. ^Steps 1-3 constitute the second meta-level of training an AI to generalize, but we can easily treat this process as a single training instance (e.g., rating how well the AI generalizes to Hindi advanced text after having been trained on doing this in 30 other languages) and iterate over and over again. I think this would look like:
    1. Running the analogs of steps 1-4 on generalizing from 
      • (a) simple text to advanced text in many languages
      • (b) easy opponents to hard ones across many games, 
      • (c) photo generation of common or general objects ("car") to rare/complex/specific ones ("interior of a 2006 Honda Accord VP"), across many classes of object
    2. And (hopefully) the system would eventually be able to generalize from simple python code training data to advanced coding tasks even though it had never seen any coding at all before this. 

And, of course, we can keep on adding piling layers on.

A few minutes of hasty Googling didn't turn up anything on this, but it seems pretty unlikely to be an original idea. But who knows! I wanted to get the idea written down and online before I had time to forget about it. 

On the off chance it hasn't been beaten thought about to death yet by people smarter than myself, I would consider together longer,  less hastily written post on the idea 

Comment by Aaron Bergman (aaronb50) on Animal welfare EA and personal dietary options · 2022-01-10T08:22:44.717Z · LW · GW

MichaelStJules is right about what I meant. While it's true that preferring not to experience something doesn't necessarily imply that the thing is net-negative, it seems to me very strong evidence in that direction. 

Comment by Aaron Bergman (aaronb50) on Animal welfare EA and personal dietary options · 2022-01-06T11:09:28.454Z · LW · GW

Entirely agree.  There are certainly chunks of my life (as a privileged first-worlder) I'd prefer not to have experienced, and these generally these seem less bad than "an average period of the same duration as a Holocaust prisoner." Given that animals are sentient, I'd put it at at ~98% that their lives are net negative.

Comment by Aaron Bergman (aaronb50) on For God's sake, Google it. · 2021-08-11T19:51:55.295Z · LW · GW

Good point; complex, real world questions/problems are often not Googleable, but I suspect a lot of time is spent dealing with mundane, relatively non-complex problems. Even in your example, I bet there is something useful to be learned from Googling "DIY cabinet instructions" or whatever. 

Comment by Aaron Bergman (aaronb50) on For God's sake, Google it. · 2021-08-11T19:49:54.885Z · LW · GW

Interesting, but I think you're way at the tail end of the distribution on this one. I bet I use Google more than 90%+ of people, but still not as much as I should.

Comment by Aaron Bergman (aaronb50) on For God's sake, Google it. · 2021-08-11T19:49:02.908Z · LW · GW

Yes - if not heretical, at least interesting to other people! I'm going to lean into the "blogging about things that seem obvious to me" thing now. 

Comment by Aaron Bergman (aaronb50) on For God's sake, Google it. · 2021-08-11T19:48:04.250Z · LW · GW

Fair enough, this might be a good counterargument though I'm very unsure. How much do mundane "brain workouts" matter? Tentatively, the lack of efficacy of brain training programs like Luminosity would suggest that they might not be doing much.

Comment by Aaron Bergman (aaronb50) on Delta variant: we should probably be re-masking · 2021-07-24T22:34:46.492Z · LW · GW

The if long COVID usually clears up after eight weeks, that would definitely weaken my point (which would be good news!) I haven’t decided if it would change my overall stance on masking though

Comment by Aaron Bergman (aaronb50) on Delta variant: we should probably be re-masking · 2021-07-24T22:33:06.761Z · LW · GW

Good point. Implicitly, I was thinking “wearing masks while indoors within ~10 feet of another person or outdoors if packed together like at a rally or concert”

Comment by Aaron Bergman (aaronb50) on Delta variant: we should probably be re-masking · 2021-07-24T22:27:41.342Z · LW · GW

I think so, thanks!

Comment by Aaron Bergman (aaronb50) on The topic is not the content · 2021-07-10T18:38:53.060Z · LW · GW

Good idea, think I will.

Comment by Aaron Bergman (aaronb50) on For reducing caffeine dependence, does daily maximum matter more, or does total daily intake matter more? · 2021-07-09T16:28:30.244Z · LW · GW

I've been wondering the exact same thing, thanks for asking.

Comment by Aaron Bergman (aaronb50) on Precognition · 2021-07-08T20:22:39.094Z · LW · GW

Thanks!

Comment by Aaron Bergman (aaronb50) on Precognition · 2021-07-07T00:35:19.772Z · LW · GW

Would you be willing to post this as a general post on the main forum? I think lots of people including myself would appreciate!

Comment by Aaron Bergman (aaronb50) on The topic is not the content · 2021-07-06T14:42:22.813Z · LW · GW

Thanks, but I have hardly any experience with Python. Need to start learning.

Comment by Aaron Bergman (aaronb50) on The topic is not the content · 2021-07-06T14:41:52.448Z · LW · GW

Yup, fixing. Gotta get better at proofreading.

Comment by Aaron Bergman (aaronb50) on The topic is not the content · 2021-07-06T14:41:26.324Z · LW · GW

From my perspective, this is why society at large needs to get better at communicating the content - so you wouldn't have to be good at "anticipating the content." 

The meaningfulness point is interesting, but I'm not sure I fully agree. Some topics can me meaningful but not interesting (high frequency trading to donate money) and visa-versa (video game design? No offense to video game designers).

Comment by Aaron Bergman (aaronb50) on Book review: Subtract: The Untapped Science of Less · 2021-05-24T22:02:28.381Z · LW · GW

By your description, it feels like the kind of book where an author picks a word and then rambles about it like an impromptu speaker. If this had an extraordinary thesis requiring extraordinary evidence like Manufacturing Consent then lots of anecdotes would make sense. But the thesis seems too vague to be extraordinary.

I get the impression of the kind of book which where a dense blogpost is stretched out to the length of a book. This is ironic for a book about subtraction.

Yup, very well-put.

Your point about anecdotes got me thinking; an "extraordinary thesis" might be conceptualized as claiming that the distribution of data significantly shifted away from some "obvious" average. If so, showing the existence of a few data points greater than, say, 4 standard deviations from the "obvious" average actually can be strong evidence in its favor. However, the same is not true for a few examples ~2 standard deviations away. Maybe Klotz's error is using anecdotes that aren't far enough away from what intuitively makes sense. 

Probably didn't explain that very well, so here is a Tweet from Spencer Greenberg making the point:

1. By Bayes Factor: suppose hypothesis “A” says a data point is nearly impossible, and hypothesis “B” says the data point is quite likely. Then the existence of that one data point (by Bayes’ rule) should move you substantially toward believing hypothesis B (relative to A).

Example: you have had a rash on your arm for 10 years (with no variability). You buy some “rash cream” off of a shady website, and within 2 hours of applying it, the rash is gone. You can be confident the cream works because it’s otherwise highly unlikely for the rash to vanish.

Comment by Aaron Bergman (aaronb50) on The case for hypocrisy · 2021-05-14T22:05:07.173Z · LW · GW

Looks like I’m in good company!

Comment by Aaron Bergman (aaronb50) on Dumb dichotomies in ethics, part 2: instrumental vs. intrinsic values · 2021-05-09T02:56:44.113Z · LW · GW

I don't think it is operationalizable, but I fail to see why 'net positive mental states' isn't a meaningful, real value. Maybe the units would be apple*minutes or something, where one unit is equivalent to the pleasure you get by eating an apple for one minute. It seems that this could in principle be calculated with full information about everyone's conscious experience. 

Comment by Aaron Bergman (aaronb50) on Dumb dichotomies in ethics, part 2: instrumental vs. intrinsic values · 2021-05-07T20:55:46.730Z · LW · GW

Are you using 'utility' in the economic context, for which a utility function is purely ordinal? Perhaps I should have used a different word, but I'm referring to 'net positive conscious mental states,' which intuitively doesn't seem to suffer from the same issues. 

Comment by Aaron Bergman (aaronb50) on On silence · 2021-04-30T23:29:42.075Z · LW · GW

Interesting, thanks. Assuming this effect is real, I wonder how much is due to the physical movement of walking rather than the low-level cognitive engagement associated with doing something mildly goal-oriented (i.e. trying to reach a destination), or something else. 

Comment by Aaron Bergman (aaronb50) on On silence · 2021-04-30T23:25:34.755Z · LW · GW

Thanks for your perspective. 

I've never been able to do intellectual work with background music, and am baffled by people e.g. programmers who work with headphones playing music all day. But maybe for them it does just use different parts of the brain

For me, there is a huge qualitative difference between lyrical music or even "interesting" classical and electronic music, and very "boring," quiet lyric-less music. Can't focus at all listening to lyrics, but soft ambient music feels intuitively helpful (though this could be illusory). This is especially the case when its a playlist or song I've heard a hundred times before, so the tune is completely unsurprising. 

Comment by Aaron Bergman (aaronb50) on What books are for: a response to "Why books don't work." · 2021-04-23T21:56:28.703Z · LW · GW

Yes, I was incorrect about Matuschak's position. He commented on reddit here:

"I think Matuschak would say that, for the purpose of conveying information, it would be much more efficient to read a very short summary than to read an entire book."

FWIW, I wouldn't say that! Actually, my research for the last couple years has been predicated on the value of embedding focused learning interactions (i.e spaced repetition prompts) into extended narrative. The underlying theory isn't (wasn't!) salience-based, but basically I believe that strong understanding is produced with a rich network of connections and a meaningful emotional connection, both of which are promoted by narrative (but usually not by a very short summary).

Comment by Aaron Bergman (aaronb50) on What books are for: a response to "Why books don't work." · 2021-04-23T04:10:55.012Z · LW · GW

Super interesting and likely worth developing into a longer post if you're so inclined. Really like this analogy.

Comment by Aaron Bergman (aaronb50) on What books are for: a response to "Why books don't work." · 2021-04-22T22:20:33.278Z · LW · GW

Great post and thanks for linking to it! Seems like books' function and utility has gotten more attention than I would have expected. 

Comment by Aaron Bergman (aaronb50) on What books are for: a response to "Why books don't work." · 2021-04-22T22:18:55.671Z · LW · GW

But then readers would have to repeat this sentence for as long as it takes to read the blog post to get the same effect. Not quite as fun.

Comment by Aaron Bergman (aaronb50) on What books are for: a response to "Why books don't work." · 2021-04-22T22:17:47.403Z · LW · GW

Yes this is an excellent point; books increase the fidelity of idea transmission because they place something like a bound on how much an idea can be misinterpreted, since one can always appeal to the author's own words (much more than a blog post or Tweet).

Comment by Aaron Bergman (aaronb50) on The secret of Wikipedia's success · 2021-04-15T11:33:08.583Z · LW · GW

It’s not that individual journalists don’t trust Wikipedia, but that they know they can’t publish an article in which a key fact comes directly from Wikipedia without any sort of corroboration. I assume, anyway. Perhaps I’m wrong.

Comment by Aaron Bergman (aaronb50) on How You Can Gain Self Control Without "Self-Control" · 2021-03-25T23:59:22.574Z · LW · GW

Great post! Is ego depletion just another way of conceptualizing rising marginal cost of effort? Like, maybe it is a fact of human psychology that the second hour of work is more difficult and unpleasant than the first. 

Comment by Aaron Bergman (aaronb50) on Nutrition for brain. · 2021-03-20T04:09:03.604Z · LW · GW

I don't know much more than you could find searching around r/nootropics, but my sense is that the relationship between diet and cognition is highly personal, so experimentation is warranted. Some do best on keto, others as a vegan, etc. With respect to particular substances, it seems that creatine might have some cognitive benefits, but once again supplementation is highly personal. DHA helps some people and induces depression in others, for example.

Also, inflammation is a common culprit/risk factor for many mental issues, so I'd expect that a generally "healthy" diet (at least, not the standard American diet), and perhaps trying an elimination diet to see if eliminating certain foods produces a marked benefit could be helpful.  Supplements like resveratrol might help with this as well. Also might be worth experimenting with fasting of different lengths; some people find that they become extremely productive after fasting for 12-14 hours. There are a million studies on all these topics that will come up in a Google search.

Comment by Aaron Bergman (aaronb50) on Direct effects matter! · 2021-03-18T03:42:35.598Z · LW · GW

Yes, you're correct. As others have correctly noted, there is no unambiguous way of determining which effects are "direct" and which are not. However, suppose decriminalization does decrease drug use. My argument emphasizes that we would need to consider the reduction in time spent enjoying drugs as a downside to decriminalization (though I doubt this would outweigh the benefits associated with lower incarceration rates). It seems to me that this point would frequently be neglected.

Comment by Aaron Bergman (aaronb50) on Nutrition for brain. · 2021-03-17T05:19:22.881Z · LW · GW

There is a good amount of this discussion at r/nootropics - of which some is evidence based and some is not. For example, see this post

Comment by Aaron Bergman (aaronb50) on Direct effects matter! · 2021-03-15T04:03:05.105Z · LW · GW

Thanks very much. Just fixed that. 

Comment by Aaron Bergman (aaronb50) on Direct effects matter! · 2021-03-14T21:17:37.588Z · LW · GW

This is a good point. Could also be that discussing only points that might impact oneself seems more credible and less dependent on empathy, even if one really does care about others directly.

Comment by Aaron Bergman (aaronb50) on Metric selection bias: why Moore's law is less important than you think · 2021-02-11T23:19:49.504Z · LW · GW

Fair point, but you'd have to think that the tendencies of the patent officers changed over time in order to foreclose that as a good metric. 

Comment by Aaron Bergman (aaronb50) on Metric selection bias: why Moore's law is less important than you think · 2021-02-11T00:20:50.548Z · LW · GW

I meant objective in the sense that the metric itself is objective, not that it is necessarily a good indicator of innovation. Yes, you're right. I do like Cowen and Southewood's method of only looking at patents registered all of the U.S., Japan, and E.U. 

Comment by Aaron Bergman (aaronb50) on Metric selection bias: why Moore's law is less important than you think · 2021-02-08T23:00:56.524Z · LW · GW

Basically agree with this suggestion: broader metrics are more likely to be unbiased over time. Even the electric grid example, though, isn't ideal because we can imagine a future point where going from $0.0001 to $0.000000001 per kilowatt-hour, for example, just isn't relevant. 

Total factor productivity and GDP per capita are even better, agreed. 

While a cop-out, my best guess is that a mixture of qualitative historical assessments (for example, asking historians, entrepreneurs, and scientists to rank decades by degree of progress) and using a variety of direct and indirect objective metrics (ex. patent rates, total factor productivity, cost of energy, life expectancy) is the best option. Any single or small group of metrics seems bound to be biased in one way or another. Unfortunately, it's hard to figure out how to weight and compare all of these things. 

Comment by Aaron Bergman (aaronb50) on Metric selection bias: why Moore's law is less important than you think · 2021-02-08T03:50:35.282Z · LW · GW

Thank you! Should have known someone would have beat me to it. 

Comment by Aaron Bergman (aaronb50) on Dumb Dichotomies in Ethics, Part 1: Intentions vs Consequences · 2021-02-06T05:00:17.609Z · LW · GW

I was thinking the third bullet, though the question of perverse incentives needs fleshing out, which I briefly alluded to at the end of the post:

“Expected consequences”, for example, leaves under-theorized when you should seek out new, relevant information to improve your forecast about some action’s consequences.

My best guess is that this isn't actually an issue, because you have a moral duty to seek out that information, as you know a priori that seeking out such info is net-positive in itself. 

Comment by Aaron Bergman (aaronb50) on Dumb Dichotomies in Ethics, Part 1: Intentions vs Consequences · 2021-01-29T23:14:03.718Z · LW · GW

Thanks for your insight. Yes, the "we simplify this for undergrads" thing seems most plausible to me. I guess my concern is that in this particular case, the simplification from "expected consequences matter" to "consequences matter" might be doing more harm than good.