AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Technology does not compel individuals to act in specific ways; rather, it presents opportunities for new forms of living and behavior. The argument suggests that while many groups might choose to ignore emerging technologies, those who adopt them often gain significant advantages. This competitive pressure can lead to a situation where others are forced to adapt to avoid falling behind. Thus, the dynamics of military and economic competition often dictate the choices societies ultimately make regarding technological adoption.
Alan Defoe serves as the Director of Frontier Safety and Governance at Google DeepMind, where he oversees three main pillars: frontier safety, governance, and planning. His team evaluates potential risks associated with emerging technologies, advising on necessary safety frameworks and regulations. The goal is to anticipate new issues that might arise with advancements in AI and ensure safe and responsible development. Collaborative efforts with various technical and policy teams facilitate a comprehensive approach to addressing these challenges.
Google DeepMind specializes in frontier models, focusing on the safety and governance of these technologies while closely collaborating with other teams within Google's broader structure. The integration of DeepMind into Google's ecosystem has fostered multi-disciplinary collaboration, enhancing the discourse on frontier issues. This collaboration aims to identify and address the policies, regulations, and safety norms that must accompany the deployment of advanced AI technologies. Another critical aspect is staying ahead of potential threats by continuously evaluating the capabilities of emerging models.
Defoe transitioned from leading the Governance of AI Center to his current role at Google DeepMind to increase his impact on AGI safety and governance. His experience in advising influential decision-makers positions him to better integrate safety considerations into real-world applications of AI. Understanding the profound implications of AI development, he emphasizes the importance of having capable leaders who are safety-oriented and aware of the stakes involved. This focus on character and competence in decision-making is crucial for addressing the complexities of AI governance.
The conversation around technological determinism highlights the intricate interplay between technology and historical development. While historical and macro-level trends influence societal outcomes, the role of human agency in steering technology continues to be debated. Some argue that societal structures and power dynamics shape technological trajectories, while others maintain the view that human decisions dictate outcomes. Understanding this relationship is vital for developing informed policies around the adoption and governance of new technologies.
Differential technological development emphasizes the need to advance beneficial technologies ahead of harmful ones to minimize risks. This approach includes developing safety measures and incentives for positive outcomes before harmful technologies become prevalent. However, challenges arise from predicting which technologies will gain traction and be adopted. A cautionary perspective suggests that rapid advancements can lead to unforeseen consequences, necessitating proactive governance frameworks to mitigate risks.
Cooperative AI seeks to enhance collaborative skills in artificial intelligence systems, fostering better interactions among them. The idea is to invest in developing systems that can work together effectively, leading to safer outcomes and reducing risks associated with autonomous decision-making. This concept extends to improving interactions between AI systems and humans, aiming to create a more harmonious coexistence. By prioritizing cooperative capabilities, researchers hope to address potential challenges posed by powerful AI technologies.
AI in education has the potential to revolutionize learning experiences by providing personalized and adaptive1-tutoring. By catering to individual needs, AI can help students who struggle or push advanced students to their full potential. This technology may also alleviate some burdens on traditional education systems, offering additional support where it's most needed. As AI systems become more integrated into classrooms, their role as educational assistants could lead to enhanced engagement and improved outcomes for students.
AI technologies like deep learning and LLMs hold the potential to transform healthcare by streamlining processes, supporting diagnosis, and increasing accessibility to expert knowledge. Innovative tools could assist doctors in patient consultations, manage treatment protocols, and expedite drug discovery through computational research. This increased efficiency and capability may lead to greater early detection of diseases and more personalized medicine. The ongoing developments in AI for health and medicine signal a promising future for better patient care.
Governance around frontier AI technologies requires collaboration across various disciplines, involving contributions from policymakers, researchers, and industry professionals. Initiatives like the Frontier Model Forum bring together key stakeholders to establish safety standards and shared governance structures. Through active engagement, these collaborative efforts aim to balance innovation and safety, ultimately leading to responsible development of AI technologies. Ensuring that societal concerns are addressed requires a concerted effort across diverse sectors.
Technology doesn’t force us to do anything — it merely opens doors. But military and economic competition pushes us through.
That’s how today’s guest Allan Dafoe — director of frontier safety and governance at Google DeepMind — explains one of the deepest patterns in technological history: once a powerful new capability becomes available, societies that adopt it tend to outcompete those that don’t. Those who resist too much can find themselves taken over or rendered irrelevant.
Links to learn more, highlights, video, and full transcript.
This dynamic played out dramatically in 1853 when US Commodore Perry sailed into Tokyo Bay with steam-powered warships that seemed magical to the Japanese, who had spent centuries deliberately limiting their technological development. With far greater military power, the US was able to force Japan to open itself to trade. Within 15 years, Japan had undergone the Meiji Restoration and transformed itself in a desperate scramble to catch up.
Today we see hints of similar pressure around artificial intelligence. Even companies, countries, and researchers deeply concerned about where AI could take us feel compelled to push ahead — worried that if they don’t, less careful actors will develop transformative AI capabilities at around the same time anyway.
But Allan argues this technological determinism isn’t absolute. While broad patterns may be inevitable, history shows we do have some ability to steer how technologies are developed, by who, and what they’re used for first.
As part of that approach, Allan has been promoting efforts to make AI more capable of sophisticated cooperation, and improving the tests Google uses to measure how well its models could do things like mislead people, hack and take control of their own servers, or spread autonomously in the wild.
As of mid-2024 they didn’t seem dangerous at all, but we’ve learned that our ability to measure these capabilities is good, but imperfect. If we don’t find the right way to ‘elicit’ an ability we can miss that it’s there.
Subsequent research from Anthropic and Redwood Research suggests there’s even a risk that future models may play dumb to avoid their goals being altered.
That has led DeepMind to a “defence in depth” approach: carefully staged deployment starting with internal testing, then trusted external testers, then limited release, then watching how models are used in the real world. By not releasing model weights, DeepMind is able to back up and add additional safeguards if experience shows they’re necessary.
But with much more powerful and general models on the way, individual company policies won’t be sufficient by themselves. Drawing on his academic research into how societies handle transformative technologies, Allan argues we need coordinated international governance that balances safety with our desire to get the massive potential benefits of AI in areas like healthcare and education as quickly as possible.
Host Rob and Allan also cover:
Chapters:
Video editing: Simon Monsour
Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
Camera operator: Jeremy Chevillotte
Transcriptions: Katy Moore
Listen to all your favourite podcasts with AI-powered features
Listen to the best highlights from the podcasts you love and dive into the full episode
Hear something you like? Tap your headphones to save it with AI-generated key takeaways
Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more
Listen to all your favourite podcasts with AI-powered features
Listen to the best highlights from the podcasts you love and dive into the full episode