Poster Session on AI Safety

post by Neil Crawford (neil-crawford) · 2022-11-12T03:50:11.528Z · LW · GW · 6 comments

6 comments

Comments sorted by top scores.

comment by Charlie Steiner · 2022-11-12T08:48:39.369Z · LW(p) · GW(p)

I agree that having a section on "what to do about it" is really useful for getting people interested. Otherwise you have a lot of unresolved tension.

Replies from: neil-crawford
comment by Neil Crawford (neil-crawford) · 2022-11-12T15:27:40.695Z · LW(p) · GW(p)

Totally! I'll make sure to include such a section next time I present on AI safety or AI governance. After a quick Google search I found the following link post which would have been useful prior to the PPE Society poster session: https://forum.effectivealtruism.org/posts/kvkv6779jk6edygug/some-ai-governance-research-ideas [EA · GW]

comment by jacob_cannell · 2022-11-12T23:49:01.208Z · LW(p) · GW(p)

Some quick comments based purely on the poster (which is probably the most important part of your funnel):

"Biological Anchors" is probably not a meaningful term for your audience.

We have a 50% chance of recreating that amount of relevant computation by 2060

This seems wrong in that we already have around brain training levels of computation [LW · GW] now or will soon - far before 2060. The remaining uncertainty is over software/algorithms, not hardware. We already have the hardware or are about to.

Once AI is capable of ML programming, it could improve its algorithms, making itself better at ML programming

This is overly specific - why only ML programming? What if the lowest hanging fruit is actually in cuda programming? Or just moving to different hardware? Or designing new hardware? Or better networking tech? Or one wierd trick to make a trillion dollars and quickly scale to more hardware? Etc etc. The idea that there are enormous gains in further optimization of ML architecture alone, and that this unending cornucopia of optimization low hanging fruit will still be bountiful and limitless by the time we actually get AGI - this suggests a very naive view of ML & neuroscience.

Just replace "ML programming" with "science and engineering R&D" or similar.

Training AI requires us to select an objective function to be maximized, yet coming up with an unproblematic objective function is really hard.

Many smart people will bounce hard off this, because they have many many examples where coming up with an unproblematic objective function isn't really hard at all. It's trivial to write the correct objective function for Chess or Go. It was trivial to design the correct utility function for atari, for minecraft even (which doesn't have a score!), it was also trivial for optimizing datacenter power usage, for generating high quality images from text, for every other modern example of DL, etc etc.

I would change this to something like:

"Training AI requires us to select an objective function to be maximized, yet coming up with an unproblematic objective function for AGI - agents with general intelligence beyond that of humans - seems really hard".

Replies from: neil-crawford
comment by Neil Crawford (neil-crawford) · 2022-11-14T06:43:53.517Z · LW(p) · GW(p)

Thanks, Jacob! This is helpful. I've made the relevant changes to my copy of the poster.

Regarding the 'biological anchors' point, I intended to capture the notion that it is not just the level/amount of computation that matters by prefixing with the word 'relevant'. When expanding on that point in conversation, I am careful to point out that generating high levels of computation isn't sufficient for creating human-level intelligence. I agree with what you say. I also think you're right about the term "biological anchors" not being very meaningful to my audience. Given that, from my experience, many academics see the poster but don't ask questions, it's probably a good idea for me to substitute this term for another. Thanks!

comment by jimv · 2022-11-12T23:22:08.383Z · LW(p) · GW(p)

I don't think you need to view namedropping as an appeal to authority. The natural way to do it in a scholarly document, including a poster, would be to cite a source. That's giving the reader valuable information - a way to check out the authority behind it.

Of course, if the reader is familiar with the author cited and knows that their work is invariably strong, they might choose to take it on authority as a shortcut, but they have the info at hand to check into it if they wish.

Replies from: neil-crawford
comment by Neil Crawford (neil-crawford) · 2022-11-14T06:54:27.927Z · LW(p) · GW(p)

I think that's right, but I think that who I cite in this case matters a lot to whether people take it seriously. This is why I chose not to cite Miles or Yudkowsky, though I'm aware that this is academically bad practice. In hindsight, I could have included some quote from Peter Railton but it doesn't feel right to do this for the sake of adding an authority to the list of citations. Thanks!