Posts

Vaguely interested in Effective Altruism? Please Take the Official 2022 EA Survey 2022-12-16T21:07:36.975Z
Please Take the 2020 Effective Altruism Survey 2020-12-07T22:18:38.213Z
Peter's COVID Consolidated Brief - 29 Apr 2020-04-29T15:27:35.638Z
Peter's COVID Consolidated Brief for 2 April 2020-04-02T17:00:06.548Z
Peter's COVID Consolidated Brief for 29 March 2020-03-29T17:07:39.085Z
Please Take the 2019 EA Survey! 2019-09-30T21:21:17.988Z
Last chance to participate in the 2018 EA Survey 2018-05-19T00:37:42.751Z
Please Take the 2018 Effective Altruism Survey! 2018-05-02T07:58:29.016Z
The 2017 Effective Altruism Survey - Please Take! 2017-04-24T21:08:54.258Z
Using a Spreadsheet to Make Good Decisions: Five Examples 2016-11-28T17:10:21.949Z
Take the EA survey, help the EA movement grow and potentially win $250 to your favorite charity 2015-12-01T17:07:56.335Z
Take the EA survey, help the EA movement grow and potentially win $250 to your favorite charity 2015-12-01T01:56:20.255Z
We’ll write you a will for free if you leave a gift to GiveWell’s top charities 2015-10-16T03:33:15.176Z
Donate to Keep Charity Science Running 2015-01-27T02:45:30.335Z
Pomodoro for Programmers 2014-12-24T18:26:49.923Z
How to Read 2014-12-22T05:40:37.235Z
You have a set amount of "weirdness points". Spend them wisely. 2014-11-27T21:09:22.804Z
Productivity 101 For Beginners 2014-11-05T23:04:46.263Z
[link] Guide on How to Learn Programming 2014-04-18T17:08:36.485Z
What has .impact done so far? 2014-03-31T17:50:19.022Z
Introducing Skillshare.im 2014-03-28T21:27:50.299Z
[LINK] Popular press account of social benefits of motivated reasoning 2014-01-12T13:55:24.241Z
Introducing .impact 2014-01-05T15:35:38.166Z
[Link] Good Judgment Project, Season Three 2013-12-02T19:51:21.982Z
Doing Important Research on Amazon's Mechanical Turk? 2013-09-25T17:04:25.692Z
Should We Tell People That Giving Makes Them Happier? 2013-09-04T21:22:06.770Z
How I Am Productive 2013-08-27T13:04:42.568Z
[LINK] "Preventing Human Extinction" by Nick Beckstead, Peter Singer, and Matt Wage 2013-08-21T11:14:41.588Z
Effective Altruist Job Board? 2013-08-18T22:40:34.868Z
Where I've Changed My Mind on My Approach to Speculative Causes 2013-08-16T07:09:54.834Z
Practical Limits to Giving Now and The Haste Consideration 2013-08-14T21:32:53.800Z
The Rebuttal Repository 2013-08-11T06:59:21.424Z
What Would it Take to "Prove" a Speculative Cause? 2013-08-07T20:59:40.037Z
How does MIRI Know it Has a Medium Probability of Success? 2013-08-01T11:42:24.140Z
Why I'm Skeptical About Unproven Causes (And You Should Be Too) 2013-07-29T09:09:27.464Z
Why Eat Less Meat? 2013-07-23T21:30:21.234Z
Introducing Effective Fundraising, a New EA Org 2013-07-21T21:39:18.945Z
For Happiness, Keep a Gratitude Journal 2013-07-15T21:11:52.203Z
College Student Philanthropy and Funding Millenium Villages 2013-06-21T14:52:41.818Z
Initial Thoughts on Personally Finding a High-Impact Career 2013-06-19T19:10:02.084Z
Giving Now Currently Seems to Beat Giving Later 2013-06-19T18:40:52.782Z
Effective Altruism Through Advertising Vegetarianism? 2013-06-12T18:50:31.353Z
[TED Talk] Peter Singer on Effective Altruism 2013-05-21T04:06:28.371Z
To Inspire People to Give, Be Public About Your Giving 2013-05-17T06:58:09.711Z
How to Build a Community 2013-05-15T05:43:05.131Z
Questions for Moral Realists 2013-02-13T05:44:44.948Z
LessWrong Survey Results: Do Ethical Theories Affect Behavior? 2012-12-19T08:40:47.861Z
A Critique of Leverage Research's Connection Theory 2012-09-20T04:28:37.011Z
Why Don't People Help Others More? 2012-08-13T23:34:43.406Z
PZ Myers on the Infeasibility of Whole Brain Emulation 2012-07-14T18:13:51.915Z

Comments

Comment by Peter Wildeford (peter_hurford) on A list of all the deadlines in Biden's Executive Order on AI · 2023-11-01T18:20:31.502Z · LW · GW

This was very helpful for you to put together - thank you!

Comment by Peter Wildeford (peter_hurford) on Alignment Grantmaking is Funding-Limited Right Now · 2023-07-28T18:12:35.464Z · LW · GW

I'd also add that historically I believe about two-thirds of LTFF's money has also come from OpenPhil, so LTFF doesn't represent a fully independent funder (though the decisionmaking around grants is pretty independent).

Comment by Peter Wildeford (peter_hurford) on New survey: 46% of Americans are concerned about extinction from AI; 69% support a six-month pause in AI development · 2023-04-05T13:28:18.367Z · LW · GW

This is a great poll and YouGov is a highly reputable pollster, but there is a significant caveat to note about the pause finding.

The way the question is framed provides information about "1000 technology leaders" who have signed a letter in favor of the pause but does not mention any opposition to the pause. I think this would push respondents to favor the pause. Ideal question construction would present more neutrally with both support and oppose statements.

Comment by Peter Wildeford (peter_hurford) on Can GPT-3 Write Contra Dances? · 2023-03-26T13:21:32.405Z · LW · GW

Good to know you won't be out of a job just yet!

Comment by Peter Wildeford (peter_hurford) on Can GPT-3 Write Contra Dances? · 2023-03-25T22:43:24.401Z · LW · GW

I asked GPT4 the same prompt and here is what it said:

Of course! Here's a new contra dance for you:

Contra Dance
Whirlwind Delight
By ChatGPT
(Duple improper)

A1


(8) Circle left 3/4


(8) Neighbor swing

 

A2


(8) Long lines forward and back


(8) Ladies chain

 

B1


(8) Star right 1x


(8) Gents aleman left 1.5

 

B2


(8) Half hey, ladies pass right shoulders to start


(8) Partner balance and swing, end facing down in a line of four

 

Enjoy dancing Whirlwind Delight!

How did it do?

Comment by Peter Wildeford (peter_hurford) on What's the Least Impressive Thing GPT-4 Won't be Able to Do · 2023-03-15T13:27:02.756Z · LW · GW

Yep! I was wrong and this is false!

Comment by Peter Wildeford (peter_hurford) on An AI risk argument that resonates with NYTimes readers · 2023-03-13T15:28:53.391Z · LW · GW

If we want to know what arguments resonate with New York Times articles we can actually use surveys, message testing, and focus groups to check and we don't need to guess! (Disclaimer: My company sells these services.)

Comment by Peter Wildeford (peter_hurford) on Let’s think about slowing down AI · 2022-12-27T17:21:30.076Z · LW · GW

Cool - I'll follow up when I'm back at work.

Comment by Peter Wildeford (peter_hurford) on Let’s think about slowing down AI · 2022-12-26T15:14:32.616Z · LW · GW

That makes a lot of sense. We can definitely test a lot of different framings. I think the problem with a lot of these kinds of problems is that they are low saliency, and thus people tend not to have opinions already, and thus they tend to generate an opinion on the spot. We have a lot of experience polling on low saliency issues though because we've done a lot of polling on animal farming policy which has similar framing effects.

Comment by Peter Wildeford (peter_hurford) on Let’s think about slowing down AI · 2022-12-25T23:07:11.399Z · LW · GW

I'll shill here and say that Rethink Priorities is pretty good at running polls of the electorate if anyone wants to know what a representative sample of Americans think about a particular issue such as this one. No need to poll Uber drivers or Twitter when you can do the real thing!

Comment by Peter Wildeford (peter_hurford) on Vaguely interested in Effective Altruism? Please Take the Official 2022 EA Survey · 2022-12-17T16:55:58.945Z · LW · GW

Yeah, it came from a lawyer. The point being that if you confess to something bad, we may be legally required to repot that, so be careful.

Comment by Peter Wildeford (peter_hurford) on Vaguely interested in Effective Altruism? Please Take the Official 2022 EA Survey · 2022-12-17T16:55:07.497Z · LW · GW

Feel free to skip questions if you feel they aren't applicable to you.

Comment by Peter Wildeford (peter_hurford) on Do anthropic considerations undercut the evolution anchor from the Bio Anchors report? · 2022-10-01T22:51:45.317Z · LW · GW

Does the chance evolution got really lucky cancel out with the chance that evolution got really unlucky? So maybe this doesn't change the mean but does increase the variance?as for how much to increase the variance, maybe like an additional +/-1 OOM tacked on to the existing evolution anchor?

I'm kinda thinking there's like a 10% chance you'd have to increase it by 10x and a 10% chance you'd have to decrease it by 10x. But maybe I'm not thinking about this right?

Comment by Peter Wildeford (peter_hurford) on No, human brains are not (much) more efficient than computers · 2022-09-06T17:49:40.525Z · LW · GW

There are a lot of different ways you can talk about "efficiency" here. The main thing I am thinking about with regard to the key question "how much FLOP would we expect transformative AI to require?" is whether, when using a neural net anchor (not evolution) to add a 1-3 OOM penalty to FLOP needs due to 2022-AI systems being less sample efficient than humans (requiring more data to produce the same capabilities) and with this penalty decreasing over time given expected algorithmic progress. The next question would be how much more efficient potential AI (e.g., 2100-AI not 2022-AI) could be given fundamentals of silicon vs. neurons, so we might know how much algorithmic progress could affect this.

I think it is pretty clear right now that 2022-AI is less sample efficient than humans. I think other forms of efficiency (e.g., power efficiency, efficiency of SGD vs. evolution) are less relevant to this.

Comment by Peter Wildeford (peter_hurford) on What's the Least Impressive Thing GPT-4 Won't be Able to Do · 2022-08-23T02:49:13.209Z · LW · GW

Yeah ok 80%. I also do concede this is a very trivial thing, not like some "gotcha look at what stupid LMs can't do no AGI until 2400".

Comment by Peter Wildeford (peter_hurford) on What's the Least Impressive Thing GPT-4 Won't be Able to Do · 2022-08-22T23:44:42.033Z · LW · GW

This is admittedly pretty trivial but I am 90% sure that if you prompt GPT4 with "Q: What is today's date?" it will not answer correctly. I think something like this would literally be the least impressive thing that GPT4 won't be able to do.

Comment by Peter Wildeford (peter_hurford) on Refine's First Blog Post Day · 2022-08-14T19:01:10.419Z · LW · GW

Thanks!

Comment by Peter Wildeford (peter_hurford) on Refine's First Blog Post Day · 2022-08-14T17:51:37.603Z · LW · GW

Is it ironic that the link to "All the posts I will never write" goes to a 404 page?

Comment by Peter Wildeford (peter_hurford) on Paper: Teaching GPT3 to express uncertainty in words · 2022-06-01T03:29:50.649Z · LW · GW

Does it get better at Metaculus forecasting?

Comment by Peter Wildeford (peter_hurford) on Call For Distillers · 2022-04-06T17:35:13.035Z · LW · GW

This sounds like something that could be done as an organization creating a job for it, which could help with mentorship/connections/motivation/job security relative to expecting people to apply to EAIF/LTFF

My organization (Rethink Priorities) is currently hiring for research assistants and research fellows (among other roles) and some of their responsibilities will include distillation.

Comment by Peter Wildeford (peter_hurford) on Late 2021 MIRI Conversations: AMA / Discussion · 2022-03-03T05:36:29.078Z · LW · GW

These conversations are great and I really admire the transparency. It's really nice to see discussions that normally happen in private happen instead in public where everyone can reflect, give feedback, and improve their own thoughts. On the other hand, the combined conversations combined to a decent-sized novel - LW says 198,846 words! Is anyone considering investing heavily in summarizing the content for people to get involved without having to read all that content?

Comment by Peter Wildeford (peter_hurford) on My experience at and around MIRI and CFAR (inspired by Zoe Curzi's writeup of experiences at Leverage) · 2021-10-17T16:34:39.347Z · LW · GW

I don't recall the specific claim, just that EY's probability mass for the claim was in the 95-99% range. The person argued that because EY disagrees with some other thoughtful people on that question, he shouldn't have such confidence.

 

I think people conflate the very reasonable "I am not going to adopt your 95-99% range because other thoughtful people disagree and I have no particular reason to trust you massively more than I trust other people" with the different "the fact that other thoughtful people mean there's no way you could arrive at 95-99% confidence" which is false. I think thoughtful people disagreeing with you is decent evidence you are wrong but can still be outweighed.

Comment by Peter Wildeford (peter_hurford) on Petrov Day 2021: Mutually Assured Destruction? · 2021-09-27T15:37:42.034Z · LW · GW

So it looks like we survived? (Yay)

Comment by Peter Wildeford (peter_hurford) on Petrov Day 2021: Mutually Assured Destruction? · 2021-09-26T21:51:14.324Z · LW · GW

I will be on the lookout for false alarms.

Comment by Peter Wildeford (peter_hurford) on Petrov Day 2021: Mutually Assured Destruction? · 2021-09-26T16:14:42.143Z · LW · GW

I can see whether the site is down or not. Seems pretty clear.

Comment by Peter Wildeford (peter_hurford) on Petrov Day 2021: Mutually Assured Destruction? · 2021-09-26T15:13:51.677Z · LW · GW

Attention LessWrong - I am a chosen user of EA Forum and I have the codes needed to destroy LessWrong. I hereby make a no first use pledge and I will not enter my codes for any reason, even if asked to do so. I also hereby pledge to second strike - if the EA Forum is taken down, I will retaliate.

Comment by Peter Wildeford (peter_hurford) on What will GPT-4 be incapable of? · 2021-04-06T23:16:39.776Z · LW · GW

Seems like "the right prompt" is doing a lot of work here. How do we know if we have given it "the right prompt"?

Do you think GPT-4 could do my taxes?

Comment by Peter Wildeford (peter_hurford) on Why Hasn't Effective Altruism Grown Since 2015? · 2021-03-09T06:53:48.524Z · LW · GW

1.) I think the core problem is that honestly no one (except 80K) actually is investing significant effort on growing the EA community since 2015 (especially comparable to the pre-2015 effort and especially as a percentage of total EA resources)

2.) Some of these examples are suspect. The GiveWell numbers definitely look to be increasing beyond 2015, especially when OpenPhil's understandably constant fundraising is removed - and this increase in GiveWell seems to line up with GiveWell's increased investment in their outreach. The OpenPhil numbers also look just to be sensitive to a few dominant eight figure grants, which understandably are not annual events. (Also my understanding is that Open Phil is starting off slowly intentionally but will aim to ramp up significantly in the near future.)

3.) As I capture in "Is EA Growing? EA Growth Metrics for 2018", many relevant EA growth statistics have peaked after 2015 or haven't peaked yet.

4.) There are still a lot of ways EA is growing other than what is captured in these graphs. For example, I bet something like total budget of EA orgs has been growing a lot even since 2015.

5.) Contrary to the "EA is inert" hypothesis, EA Survey data has shown that many people have been "convinced" of EA. Furthermore, our general population surveys show that the vast majority of people (>95% of US university students) have still not heard of EA.

Comment by Peter Wildeford (peter_hurford) on Why Hasn't Effective Altruism Grown Since 2015? · 2021-03-09T06:44:13.027Z · LW · GW

FWIW I I put together "Is EA Growing? EA Growth Metrics for 2018" and I'm looking forward for doing 2019+2020 soon

Comment by Peter Wildeford (peter_hurford) on What gripes do you have with Mustachianism? · 2020-06-12T15:53:10.934Z · LW · GW

Mr. Money Mustache has a lot of really good advice that I find a lot of value from. However, I think Mr. Money Mustache underestimates the ease and impact of opportunities to grow income relative to cutting spending - especially if you're in (or can be in) a high-earning field like tech. Doubling your income will put you on a much faster path than cutting your spending a further 5%.

Comment by Peter Wildeford (peter_hurford) on What are the best tools for recording predictions? · 2020-05-25T15:21:08.491Z · LW · GW

PredictionBook is really great for lightweight, private predictions and does everything you're looking for. Metaculus is great for more fully-featured predicting and I believe also supports private questions, but may be a bit of overkill for your use case. A spreadsheet also seems more than sufficient, as others have mentioned.

Comment by Peter Wildeford (peter_hurford) on Peter's COVID Consolidated Brief - 29 Apr · 2020-04-29T19:48:37.564Z · LW · GW

Thanks. I'll definitely aim to produce them more quickly... this one got away from me.

Comment by Peter Wildeford (peter_hurford) on Coronavirus: the four levels of social distancing, and when and why we might transition between them · 2020-03-28T04:07:28.973Z · LW · GW

My understanding is that we also have and might in the future also spend a decent amount of time in a "level 2.5", where some but not all non-essential businesses are open (i.e., no groups larger than ten, restaurants are closed to dine-in, hair salons are open).

Comment by Peter Wildeford (peter_hurford) on SARS-CoV-2 pool-testing algorithm puzzle · 2020-03-21T13:56:09.382Z · LW · GW

A binary search strategy still could be more efficient, depending on the ratio of positives to negatives.

Comment by Peter Wildeford (peter_hurford) on SARS-CoV-2 pool-testing algorithm puzzle · 2020-03-21T04:38:09.360Z · LW · GW

What about binary search?

Comment by Peter Wildeford (peter_hurford) on Is the coronavirus the most important thing to be focusing on right now? · 2020-03-21T01:22:08.308Z · LW · GW

This is a good answer.

Comment by Peter Wildeford (peter_hurford) on Is the coronavirus the most important thing to be focusing on right now? · 2020-03-21T01:19:15.642Z · LW · GW

Not really an answer, but a statement and a question - I imagine this is literally the least neglected issue in the world right now. How much does that affect the calculus? How much should we defer to people with more domain expertise?

Comment by Peter Wildeford (peter_hurford) on What is your recommended statistics textbook for a beginner? · 2019-12-29T17:27:34.791Z · LW · GW

Introduction to Statistical Learning

Comment by Peter Wildeford (peter_hurford) on Understanding “Deep Double Descent” · 2019-12-06T14:58:49.376Z · LW · GW

This paper also seems helpful: https://arxiv.org/pdf/1812.11118.pdf

Comment by Peter Wildeford (peter_hurford) on Please Take the 2019 EA Survey! · 2019-10-18T19:53:28.592Z · LW · GW

Answered here: https://forum.effectivealtruism.org/posts/YAwLfgwhg7opp3rTp/please-take-the-2019-ea-survey#G8Hn64AEyh3uMY2SG

Comment by Peter Wildeford (peter_hurford) on Please Take the 2019 EA Survey! · 2019-10-15T16:53:16.234Z · LW · GW

The EA Survey is closing today! Please take! https://www.surveymonkey.co.uk/r/EAS2019LW

Comment by Peter Wildeford (peter_hurford) on Please Take the 2019 EA Survey! · 2019-10-01T01:43:13.622Z · LW · GW

Thanks!

Comment by Peter Wildeford (peter_hurford) on Honoring Petrov Day on LessWrong, in 2019 · 2019-09-26T21:37:08.972Z · LW · GW

It could also be on the list of pros, depending on how one uses LW.

Comment by Peter Wildeford (peter_hurford) on Honoring Petrov Day on LessWrong, in 2019 · 2019-09-26T18:08:39.989Z · LW · GW

Are you offering to take donations in exchange for pressing the button or not pressing the button?

Comment by Peter Wildeford (peter_hurford) on The why and how of daily updates · 2019-05-06T03:47:14.209Z · LW · GW

What happens if you don't check off everything for the day?

Comment by peter_hurford on [deleted post] 2018-07-10T05:42:38.432Z

This sounds fairly similar to being on a board of a non-profit.

Comment by Peter Wildeford (peter_hurford) on Success and Fail Rates of Monthly Policies · 2017-12-11T07:56:17.580Z · LW · GW

Nice post. I'd be curious to hear what all the monthly themes were.

Comment by Peter Wildeford (peter_hurford) on My simple hack for increased alertness and improved cognitive functioning: very bright light · 2017-09-26T03:22:51.804Z · LW · GW

Does anyone know if there are links to purchase some of this online?

Comment by Peter Wildeford (peter_hurford) on Voting Weight Discussion · 2017-09-26T02:15:29.236Z · LW · GW

I don't think so. The second equation is negative infinity for karma = 0, which seems not right.

Comment by Peter Wildeford (peter_hurford) on Voting Weight Discussion · 2017-09-26T02:13:47.055Z · LW · GW

That makes sense. On a mostly unrelated note, is there any way to get notified when someone replies to my comment?