Podcast interview series featuring Dr. Peter Park
post by jacobhaimes · 2024-03-26T00:25:58.129Z · LW · GW · 0 commentsThis is a link post for https://into-ai-safety.github.io/episode/interview/episode-17/
Contents
Episode 1 | StakeOut.AI Milestones Episode 2 | The Next AI Battlegrounds Episode 3 | Freeform Acknowledgements None No comments
Check out the Into AI Safety podcast on Spotify, Apple Podcasts, Amazon Music, YouTube Podcasts, and many other podcast listening platforms!
As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.
The sub-series runs approximately 3 hours and 40 minutes in total, during which Dr. Park and I discuss StakeOut.AI, a nonprofit which he cofounded along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position.
The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.
If you would like to investigate further into Dr. Park’s work, view his website, Google Scholar, or follow him on Twitter!
Since the interview is so long, I totally get wanting to jump right to the parts you are most interested in. To assist with this, I have included chapter timestamps in the show notes, which should allow you to quickly find the content you're looking for. In addition, will give a brief overview of each episode here, without going into too much detail. You can find even more sources on the Into AI Safety website.
Episode 1 | StakeOut.AI Milestones
- Milestones
- Last minute lobbying to water down the EU AI Act
- Divide-and-Conquer Dynamics in AI-Driven Disempowerment [7]
- AI "art"
Episode 2 | The Next AI Battlegrounds
- Battlegrounds
- OpenAI Drama
- Corporate influence campaigns
Episode 3 | Freeform
- The power grab of rapid development
- Provable safety
- The great Open Source debate
- AIxBio and scientific rigor in AI
- Deception about AI deception [21]
- "I'm sold, next steps?" -you
Acknowledgements
This work was made possible by AI Safety Camp
Special thanks to individuals that helped along the way:
Dr. Peter Park; Chase Precopia; Brian Penny; Leah Selman; Remmelt Ellen; Pete Wright
0 comments
Comments sorted by top scores.