The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Sam Charrington
undefined
93 snips
Dec 17, 2025 • 53min

Rethinking Pre-Training for Agentic AI with Aakanksha Chowdhery - #759

Aakanksha Chowdhery, a machine learning researcher from Reflection, dives into the future of agentic AI. She critiques the current reliance on post-training techniques, advocating for a transformative approach to pre-training. Aakanksha highlights the need for evolving attention mechanisms and tailored training data to enhance long-term reasoning and planning. She discusses the importance of 'trajectory' training data and the risks of synthetic data, all while underscoring the significance of rigorous evaluation in building agentic models.
undefined
67 snips
Dec 9, 2025 • 58min

Why Vision Language Models Ignore What They See with Munawar Hayat - #758

Munawar Hayat, a researcher at Qualcomm AI Research specializing in multimodal generative AI, dives into the intricacies of Vision-Language Models (VLMs). He discusses the puzzling issue of object hallucination, revealing why these models often overlook visual elements in favor of language. Munawar also introduces attention-guided alignment techniques and a novel approach to generalized contrastive learning for efficient multi-modal retrieval. He shares insights on the Multi-Human Testbench designed to tackle identity leakage challenges in generative models, bringing clarity to this evolving field.
undefined
78 snips
Dec 2, 2025 • 49min

Scaling Agentic Inference Across Heterogeneous Compute with Zain Asgar - #757

Zain Asgar, co-founder and CEO of Gimlet Labs, is an expert in efficient AI compute orchestration and heterogeneous inference. He discusses the challenges of handling token-heavy agentic workloads and the need for diverse hardware solutions. Zain elaborates on Gimlet's innovative three-layer architecture for workload disaggregation and LLM-driven optimization. He shares insights on the complexities of networking, the trade-offs in precision, and the future of resource scheduling, all while emphasizing the importance of cost-effective AI operations.
undefined
140 snips
Nov 19, 2025 • 56min

Proactive Agents for the Web with Devi Parikh - #756

Join Devi Parikh, co-founder of Yutori and a pioneering AI researcher, as she discusses the futuristic concept of proactive web agents that can streamline our online interactions. Devi explains how these agents operate using visual models that leverage screenshots for improved reliability over traditional methods. She reveals insights into Yutori's innovative training process, the challenges of web navigation, and the pivotal role of feedback in enhancing agent performance. Tune in to explore how these advancements could reshape our digital experiences!
undefined
128 snips
Nov 12, 2025 • 55min

AI Orchestration for Smart Cities and the Enterprise with Robin Braun and Luke Norris - #755

Robin Braun, VP of AI at HPE, and Luke Norris, CEO of Kamiwaza, delve into AI's role in transforming smart cities and enterprises. They explore the 'Agentic Smart City' project in Vail, focusing on automating website accessibility and digitizing legacy data. The discussion highlights the significance of private cloud infrastructure, fresh data, and practical use cases that deliver real ROI. Learn how innovative technologies can enhance fire detection and urban monitoring, reducing costs and improving compliance in complex workflows.
undefined
103 snips
Nov 4, 2025 • 56min

Building an AI Mathematician with Carina Hong - #754

Carina Hong, founder and CEO of Axiom, discusses her groundbreaking work on creating an AI mathematician. She highlights the convergence of advanced reasoning capabilities, formal proof languages like Lean, and breakthroughs in code generation as key trends revitalizing AI in mathematics. Carina delves into the challenges of translating informal proofs into machine-verifiable formats, the concept of self-improvement through conjecturing and proving, and the potential applications in high-stakes software verification.
undefined
69 snips
Oct 28, 2025 • 52min

High-Efficiency Diffusion Models for On-Device Image Generation and Editing with Hung Bui - #753

Hung Bui, Technology VP at Qualcomm and former head of VinAI Research, dives into high-efficiency generative AI models. He discusses the transition from complex GANs to streamlined diffusion models, revealing his team's breakthroughs with SwiftBrush and SwiftEdit, which allow rapid text-to-image generation and editing. Hung explains their innovative distillation framework that creates a single-step model from a multi-step teacher, and how this paves the way for real-time, on-device AI. The conversation also touches on privacy in personalized agents and efficiency challenges in AI deployment.
undefined
167 snips
Oct 22, 2025 • 1h 13min

Vibe Coding's Uncanny Valley with Alexandre Pesant - #752

In this conversation, Alexandre Pesant, AI lead at Lovable, delves into vibe coding, a transformative practice in software development. He explains how this approach shifts focus from typing code to expressing intent through higher-level abstractions. Alex shares insights on coding agents and the critical role of context engineering. He discusses Lovable’s evolution from complex agent systems to simpler workflows and back as technologies improved. Listeners also gain perspectives on AI's impact on engineering roles and the importance of robust evaluations in creating successful AI-native development tools.
undefined
58 snips
Oct 14, 2025 • 58min

Dataflow Computing for AI Inference with Kunle Olukotun - #751

Kunle Olukotun, a Stanford professor and chief technologist at SambaNova Systems, dives into reconfigurable dataflow architectures for AI. He explains how this innovative approach enhances AI inference by dynamically matching hardware to model dataflows, leading to reduced bandwidth bottlenecks. Kunle also highlights the benefits of fast model switching and efficient multi-model serving, crucial for low-latency applications. Plus, he explores future possibilities of using AI to create compilers for evolving hardware setups, offering insights into significant performance improvements.
undefined
205 snips
Oct 7, 2025 • 57min

Recurrence and Attention for Long-Context Transformers with Jacob Buckman - #750

Jacob Buckman, co-founder and CEO of Manifest AI, dives deep into the world of long-context transformers. He discusses innovative techniques like windowed attention and the revolutionary Power Retention approach, which melds attention and recurrence for astonishing training speeds. Buckman also shares insights on Manifest AI's open-source tools, Vidrial and PowerCoder, and explores the significance of metrics in measuring context utility. Learn about the balance between state and weight for optimal compute architectures and the future potential of context length in AI.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app