AI Safety…Ok Doomer: with Anca Dragan
Building safe and capable models is one of the greatest challenges of our time. Can we make AI work for everyone? How do we prevent existential threats? Why is alignment so important? Join Professor Hannah Fry as she delves into these critical questions with Anca Dragan, lead for AI safety and alignment at Google DeepMind.
Want to share feedback? Have a suggestion for a guest that we should have on next? Why not leave a review on YouTube and stay tuned for future episodes.
Timecodes:
00:00 Introduction to Anca Dragan
02:16 Short and long term risks
04:35 Designing a safe bridge
05:36 Robotics
06:56 Human and AI interaction
12:33 The objective of alignment
14:30 Value alignment and recommendation systems
17:57 Ways to approach alignment with competing objectives
19:54 Deliberate alignment
22:24 Scalable oversight
23:33 Example of scalable oversight
26:14 What comes next?
27:20 Gemini
30:14 Long term risk and frontier safety framework
35:09 Importance of AI safety
38:02 Conclusion
Further reading:
https://deepmind.google/discover/blog/introducing-the-frontier-safety-framework/
https://arxiv.org/pdf/2403.13793
Thanks to everyone who made this possible, including but not limited to:
Presenter: Professor Hannah Fry
Series Producer: Dan Hardoon
Editor: Rami Tzabar, TellTale Studios
Commissioner & Producer: Emma Yousif
Music composition: Eleni Shaw
Camera Director and Video Editor: Tommy Bruce
Audio Engineer: Perry Rogantin
Video Studio Production: Nicholas Duke
Video Editor: Bilal Merhi
Video Production Design: James Barton
Visual Identity and Design: Eleanor Tomlinson
Commissioned by Google DeepMind
Want to share feedback? Have a suggestion for a guest that we should have on next? Why not leave a review on YouTube and stay tuned for future episodes.
Timecodes:
00:00 Introduction to Anca Dragan
02:16 Short and long term risks
04:35 Designing a safe bridge
05:36 Robotics
06:56 Human and AI interaction
12:33 The objective of alignment
14:30 Value alignment and recommendation systems
17:57 Ways to approach alignment with competing objectives
19:54 Deliberate alignment
22:24 Scalable oversight
23:33 Example of scalable oversight
26:14 What comes next?
27:20 Gemini
30:14 Long term risk and frontier safety framework
35:09 Importance of AI safety
38:02 Conclusion
Further reading:
https://deepmind.google/discover/blog/introducing-the-frontier-safety-framework/
https://arxiv.org/pdf/2403.13793
Thanks to everyone who made this possible, including but not limited to:
Presenter: Professor Hannah Fry
Series Producer: Dan Hardoon
Editor: Rami Tzabar, TellTale Studios
Commissioner & Producer: Emma Yousif
Music composition: Eleni Shaw
Camera Director and Video Editor: Tommy Bruce
Audio Engineer: Perry Rogantin
Video Studio Production: Nicholas Duke
Video Editor: Bilal Merhi
Video Production Design: James Barton
Visual Identity and Design: Eleanor Tomlinson
Commissioned by Google DeepMind
Google DeepMind
Artificial intelligence could be one of humanity's most useful inventions. DeepMind aims to build advanced AI to expand our knowledge and find new answers. By solving this one thing, we believe we could help people solve thousands of problems.
We’re a te...