Posts

Announcing aisafety.training 2023-01-21T01:01:40.580Z
Implication of AI timelines on planning and solutions 2021-08-21T05:12:12.006Z

Comments

Comment by JJ Hepburn (jj-hepburn) on How MATS addresses “mass movement building” concerns · 2023-05-05T00:30:08.315Z · LW · GW

Recently, many AI safety movement-building programs have been criticized for attempting to grow the field too rapidly and thus:

Can you link to these?

Comment by JJ Hepburn (jj-hepburn) on What is it like doing AI safety work? · 2023-02-22T01:23:54.328Z · LW · GW

This is great! Thanks for doing this.

Would you be able to add people's titles and affiliations for some context? Possibly also links to their websites, LinkedIn or similar.



 

Comment by JJ Hepburn (jj-hepburn) on Announcing aisafety.training · 2023-02-20T23:56:40.256Z · LW · GW

You can now also subscribe to be automatically emailed when new events are added or updated. You can opt for either daily or weekly updates 
 

Signup here:

https://airtable.com/shrEp75QWoCrZngXg

Comment by JJ Hepburn (jj-hepburn) on The metaphor you want is "color blindness," not "blind spot." · 2022-02-14T05:59:37.621Z · LW · GW

I have always thought of it like a vehicle blind spot not an ocular blind spot. More related to the structure of the situation than the individual.

Comment by JJ Hepburn (jj-hepburn) on What questions do you have about doing work on AI safety? · 2021-12-24T12:21:54.974Z · LW · GW

How many places did you apply for before getting your current role or position?

How much time have you spent on applying for open opportunities?

What are some things that your org has that others don’t and should?

What are some things that other orgs have that your org should have?

Comment by JJ Hepburn (jj-hepburn) on What questions do you have about doing work on AI safety? · 2021-12-22T13:12:57.587Z · LW · GW

What are some boring parts of your job that you have to do?

What are some frustrating parts of your job that you have to do?

What aspects of your job/place of work are different from what you expected from the outside?

Do you feel like you have good job security?

Comment by JJ Hepburn (jj-hepburn) on Interviews on Improving the AI Safety Pipeline · 2021-12-09T04:46:43.704Z · LW · GW

Not exactly sure what I was trying to say here. Probably using the PhD as an example of a  path to credentials. 

Here are some related things I believe:

  • I don't think a PhD is necessary or the only way
  • University credentials are not now and should not be the filter for people working on these problems
  • There is often a gap between peoples competencies and their ability to signal them
  • Credentials are the default signal for competence 
  • Universities are incredibly inefficient ways to gain competence or signal
  • Assessing people is expensive and so reviewers are incentivised to find cheaper to assess signals
  • Credentials are used as signals not because they are good but because they are cheap to assess and universally understood
  • Credentials are often necessary but rarely sufficient
Comment by JJ Hepburn (jj-hepburn) on Visible Thoughts Project and Bounty Announcement · 2021-12-01T04:10:34.995Z · LW · GW

Could do Go, Poker or some E-Sports with commentary. Poker unlike chess has the advantage that the commentators can see all of the players hands but the players can only see their own. Commentators often will talk about what a player must be thinking in this situation and account for what is observable to the player or not.

This would certainly be easier to scale but not as good quality.

Comment by JJ Hepburn (jj-hepburn) on Implication of AI timelines on planning and solutions · 2021-08-21T15:48:03.011Z · LW · GW

The plan and numbers I lay out above you actually finish friendly AI in 2036, which is the 10% point

Comment by JJ Hepburn (jj-hepburn) on Implication of AI timelines on planning and solutions · 2021-08-21T13:27:29.471Z · LW · GW

Yes, if you have a solution in 2026 it isn't likely to be relevant to something used in 2050. But 2026 is the planned solution date and 2050 is the median TAI date. 

The numbers I used above a just to demonstrate the point thought. The broad idea is that coming up with a solution/theory to alignment takes longer than planned. Having a theory isn't enough, you still have some time to make it count. Then TAI might come at the early end of your probability distribution. 

It's pretty optimistic to plan that TAI will come at your median estimate and that you won't run into the planning fallacy.

Comment by JJ Hepburn (jj-hepburn) on Raising funds to establish a new AI Safety charity · 2018-03-17T12:32:07.197Z · LW · GW

Really excited about this! Donation on the way