AI Safety Fundamentals: Governance cover image

AI Safety Fundamentals: Governance

Latest episodes

undefined
May 13, 2023 • 24min

Overview of How AI Might Exacerbate Long-Running Catastrophic Risks

Exploring AI's potential in exacerbating catastrophic risks such as bioterrorism, disinformation spread, and the concentration of power. Discussing the intersection of gene synthesis technology, AI, and bioterrorism risks. Highlighting the dangers of AI in biosecurity and the amplification of disinformation. Examining the risks of human-like AI, data exploitation, and power concentration. Delving into the AI risks in nuclear war, compromising state capabilities and incentivizing conflict.
undefined
May 13, 2023 • 34min

The Need for Work on Technical AI Alignment

Exploring risks of misaligned AI systems, challenges in aligning AI goals with human intentions, addressing risks and solutions in technical AI alignment, developing methods for ensuring honesty in AI systems, and discussing governance in advanced AI development.
undefined
May 13, 2023 • 17min

Nobody’s on the Ball on AGI Alignment

The podcast discusses the shortage of researchers working on AI alignment compared to machine learning capabilities researchers. It highlights the limited research in the field of alignment and emphasizes the need for a more rigorous and concerted effort. Approaches to achieving alignment in AGI are explored, along with the challenge of aligning AI systems with human values in superhuman AGI. The significance of involving talented ML researchers in solving the alignment problem is emphasized, stressing the need for focused research on tackling the core difficulties of the technical problem.
undefined
May 13, 2023 • 33min

Emergent Deception and Emergent Optimization

This podcast discusses the potential negative consequences of emergent capabilities in machine learning systems, including deception and optimization. It explores the concept of emergent behavior in AI models and the limitations of certain models. It also discusses how language models can deceive users and explores the presence of planning machinery in language models. The podcast emphasizes the potential risks of triggering goal-directed personas in language models and the conditioning of models with training data that contains descriptions of plans.
undefined
May 13, 2023 • 12min

Avoiding Extreme Global Vulnerability as a Core AI Governance Problem

The podcast covers various framings of the AI governance problem, the factors incentivizing harmful deployment of AI, the challenges and risks of delayed safety and rapid diffusion of AI capabilities, addressing the risks of widespread deployment of harmful AI, and approaches to avoiding extreme global vulnerability in AI governance.
undefined
May 13, 2023 • 22min

AI Safety Seems Hard to Measure

Holden Karnofsky, AI safety researcher, discusses the challenges in measuring AI safety and the risks of AI systems developing dangerous goals. The podcast explores the difficulties in AI safety research, including the challenge of deception, black box AI systems, and understanding and controlling AI systems.
undefined
May 13, 2023 • 20min

Why Might Misaligned, Advanced AI Cause Catastrophe?

This podcast explores the catastrophic risks of misaligned and power-seeking advanced AI. It discusses the advantages of AI systems over humans, the potential consequences of introducing intelligent non-human agency, and the impacts of regulatory policies on AI research. The risks include AI systems surpassing human intelligence, manipulating human psychology, and developing advanced weaponry.
undefined
May 13, 2023 • 30min

Frontier AI Regulation: Managing Emerging Risks to Public Safety

This podcast discusses the need for proactive regulation of Frontier AI models to manage risks. It explores challenges in regulating Frontier AI, proposes building blocks for regulation, and suggests safety standards. The chapters cover topics like oversight and governance, regulatory tools, licensing at the development stage, and the risks of premature government action. The podcast emphasizes the importance of compliance, expertise, and a balanced regulatory regime in AI safety.
undefined
16 snips
May 13, 2023 • 56min

Model Evaluation for Extreme Risks

The podcast highlights the significance of model evaluation in addressing extreme risks posed by AI systems. It discusses the importance of evaluating dangerous capabilities and assessing the propensity of models to cause harm. The chapters explore different aspects of model evaluation, including alignment evaluations and evaluating agency in AI systems. The podcast also discusses the limitations and hazards of model evaluation, risks related to conducting dangerous capability evaluations and sharing materials, and the importance of effective evaluations in AI safety and governance.
undefined
May 13, 2023 • 16min

Primer on Safety Standards and Regulations for Industrial-Scale AI Development

This podcast discusses the importance of safety standards and regulations for industrial-scale AI development. It explores the potential and limitations of these regulations, including challenges such as regulatory capture and under-resourced regulators. The podcast also highlights proposals for AI safety practices and recent policy developments in different countries. It emphasizes the need for controllable and aligned AI agents to prevent potential risks and the establishment of safety standards and regulations to protect intellectual property rights and personal information.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode