The Information Bottleneck

Ravid Shwartz-Ziv & Allen Roush
undefined
Jan 7, 2026 • 1h 12min

EP21: Privacy in the Age of Agents with Niloofar Mireshghallah

Niloofar Mireshghallah, an incoming assistant professor at Carnegie Mellon University, dives into the intriguing world of AI privacy and model behavior. She discusses the surprising reliance of models on context over memorization and highlights modern privacy threats like aggregation and inference attacks. The conversation touches on linguistic colonialism in AI, the challenges faced by non-English languages, and the importance of academic research in preserving the nuances of learning and cultural representation. Niloofar calls for innovative AI tools for science and education while emphasizing the need for privacy-aware designs.
undefined
35 snips
Dec 15, 2025 • 1h 50min

EP20: Yann LeCun

Yann LeCun, a Turing Award-winning computer scientist and pioneer of deep learning, shares his bold vision for AI after leaving Meta to start Advanced Machine Intelligence. He critiques the current Silicon Valley obsession with scaling language models, arguing they won't lead to artificial general intelligence. Instead, he advocates for developing world models that simulate abstract concepts. Yann discusses learning object permanence and the challenges of game AI, while advocating for safety measures in AI design, emphasizing an architecture-focused approach.
undefined
Dec 10, 2025 • 1h 11min

EP19: AI in Finance and Symbolic AI with Atlas Wang

Atlas Wang, a faculty member at UT Austin and research director at XTX Research, dives into the captivating realms of symbolic AI and quantitative finance. He unveils how neural networks can learn symbolic equations through gradient descent, challenging our understanding of AI reasoning. In finance, he demystifies high-frequency trading’s complexities, highlighting the struggle against market noise and the pursuit of predictive accuracy. Atlas advocates for robust time-series models and suggests that researchers should explore finance for the unique opportunities it presents.
undefined
Dec 1, 2025 • 1h 45min

EP18: AI Robotics

In this episode, we hosted Judah Goldfeder, a PhD candidate at Columbia University and student researcher at Google, to discuss robotics, reproducibility in ML, and smart buildings.Key topics covered:Robotics challenges: We discussed why robotics remains harder than many expected, compared to LLMs. The real world is unpredictable and unforgiving, and mistakes have physical consequences. Sim-to-real transfer remains a major bottleneck because simulators are tedious to configure accurately for each robot and environment. Unlike text, robotics lacks foundation models, partly due to limited clean, annotated datasets and the difficulty of collecting diverse real-world data.Reproducibility crisis: We discussed how self-reported benchmarks can lead to p-hacking and irreproducible results. Centralized evaluation systems (such as Kaggle or ImageNet challenges), where researchers submit algorithms for testing on hidden test sets, seem to drive faster progress.Smart buildings: Judah's work at Google focuses on using ML to optimize HVAC systems, potentially reducing energy costs and carbon emissions significantly. The challenge is that every building is different. It makes the simulation configuration extremely labor-intensive. Generative AI could help by automating the process of converting floor plans or images into accurate building simulations.Links:Judah website - https://judahgoldfeder.com/Music:"Kid Kodi" — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0."Palms Down" — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.Changes: trimmed
undefined
Nov 24, 2025 • 1h 6min

EP17: RL with Will Brown

In this conversation with Will Brown, research lead at Prime Intellect specializing in reinforcement learning (RL) and multi-agent systems, they explore the foundations and practical applications of RL. Will shares insights into the challenges RL faces in LLMs, emphasizing the importance of online sampling and reward models. He discusses multi-agent dynamics, optimization techniques, and the role of game theory in AI development. The discussion also highlights the significance of intermediate results and the future directions for RL in various applications.
undefined
Nov 17, 2025 • 59min

EP16: AI News and Papers

Dive into the intriguing world of AI as the hosts dissect the quirks of conference review dynamics and the innovation of Kimi K2 thinking. Discover Google's latest TPU advancements and their competition with NVIDIA. The importance of real-world data in robotics takes center stage, while the concept of Chain of Thought Hijacking raises alarms about model vulnerabilities. Simplifying reinforcement learning with JustRL presents new possibilities, and the Cosmos project sparks curiosity around AI-driven scientific discovery.
undefined
5 snips
Nov 13, 2025 • 1h 23min

EP15: The Information Bottleneck and Scaling Laws with Alex Alemi

In this discussion, Alex Alemi—a prominent AI researcher from Anthropic, formerly with Google Brain and Disney—delves into the concept of the information bottleneck. He explains how it captures the essential aspects of data while avoiding overfitting. Alemi also highlights scaling laws, revealing how smaller experiments can forecast larger behaviors in AI. He offers insights on the importance of compression in understanding models and challenges researchers to pursue ambitious questions that address broader implications for society, such as job disruption.
undefined
Nov 10, 2025 • 57min

EP14: AI News and Papers

Explore the intricate dynamics of AI in healthcare, where GPT-5's fragility raises pressing concerns about reliability and education. Discover Stanford's innovative Cartridges approach, which aims to cut down the hefty computational costs of long-context models. Delve into the transformative potential of Continuous Autoregressive Language Models and learn about practical strategies from the Smol Training Playbook. Join in the debate on benchmarks and dataset challenges in this fast-evolving field!
undefined
Nov 7, 2025 • 1h 21min

EP13: Recurrent-Depth Models and Latent Reasoning with Jonas Geiping

Jonas Geiping, a machine learning researcher at the ELLIS Institute and Max Planck Institute, explores the fascinating world of recurrent-depth models and latent reasoning. He discusses how these models can enhance AI's reasoning capabilities, especially in complex tasks like math and coding. The conversation also delves into challenges in model development, the importance of interpretability and safety in AI, and the future of scalable algorithms. With practical advice for budding researchers, Jonas sheds light on Tübingen as an emerging hub for machine learning innovation.
undefined
Nov 3, 2025 • 58min

EP12: Adversarial attacks and compression with Jack Morris

Join Jack Morris, a PhD student at Cornell and creator of the TextAttack library, as he dives into the intriguing world of adversarial examples in language models. Jack discusses the evolution of TextAttack, the complexities of open-source AI, and the security implications of inversion attacks. He highlights the Platonic representation hypothesis and its impact on understanding model embeddings. Plus, he connects the dots between compression in language models and their efficiency. Get ready for a fascinating exploration of the future of AI!

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app