
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.
Latest episodes

10 snips
Mar 25, 2024 • 48min
V-JEPA, AI Reasoning from a Non-Generative Architecture with Mido Assran - #677
Join Mido Assran, a research scientist at Meta's FAIR, as he delves into the groundbreaking V-JEPA model, which aims to bridge human and machine intelligence. He explains how V-JEPA's self-supervised training enables efficient learning from unlabeled video data without the distraction of pixel details. Mido also tackles innovations in visual prediction, the use of advanced techniques for video processing, and the complexities of temporal prediction. This insightful conversation highlights the future of AI reasoning beyond generative models.

18 snips
Mar 18, 2024 • 50min
Video as a Universal Interface for AI Reasoning with Sherry Yang - #676
Sherry Yang, a Senior Research Scientist at Google DeepMind and a PhD candidate at UC Berkeley, discusses her groundbreaking work on video as a universal interface for AI reasoning. She draws parallels between video generation models and language models, highlighting their potential in real-world decision-making tasks. The conversation covers the integration of video in robotics, the challenges of effective labeling, and the exciting applications of interactive simulators. Sherry also unveils UniSim, showcasing the future of engaging with AI-generated environments.

5 snips
Mar 11, 2024 • 40min
Assessing the Risks of Open AI Models with Sayash Kapoor - #675
Sayash Kapoor, a Ph.D. student at Princeton University, discusses his research on the societal impact of open foundation models. He highlights the controversies surrounding AI safety and the potential risks of releasing model weights. The conversation delves into critical issues, such as biosecurity concerns linked to language models and the challenges of non-consensual imagery in AI. Kapoor advocates for a unified framework to evaluate these risks, emphasizing the need for transparency and legal protections in AI development.

29 snips
Mar 4, 2024 • 32min
OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia - #674
Akshita Bhagia, a senior research engineer at the Allen Institute for AI, shares her insights on OLMo, an open-source language model that includes a unique dataset and tools for training. She discusses the innovative Dolma dataset, which boasts a three-trillion-token corpus, and Paloma, a benchmarking tool for evaluating model performance. Throughout the conversation, Akshita emphasizes the importance of data transparency, collaborative research, and the challenges faced in training large-scale models, advocating for a shared knowledge approach in AI development.

6 snips
Feb 26, 2024 • 25min
Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673
Ben Prystawski, a PhD student at Stanford blending cognitive science with machine learning, unveils fascinating insights on LLM reasoning. He discusses his recent paper that questions if reasoning exists in LLMs and the effectiveness of chain-of-thought strategies. Delve into how locality in training data fuels reasoning capabilities, and explore the nuances of optimizing prompts for better model performance. The conversation also touches on how our human experiences shape reasoning, enhancing comprehension in artificial intelligence.

24 snips
Feb 19, 2024 • 46min
Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672
Armineh Nourbakhsh, Executive Director at JP Morgan AI Research, dives into the exciting world of DocLLM, a layout-aware large language model designed for document understanding. She shares insights on the evolution of document AI, focusing on multimodal approaches that combine textual and visual data. Nourbakhsh discusses the challenges of training generative models, the intricacies of processing enterprise documents, and strategies to reduce hallucinations in language models, enhancing performance in complex document analysis.

23 snips
Feb 12, 2024 • 1h 6min
Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671
Sanmi Koyejo, an assistant professor at Stanford University, dives into the fascinating world of large language models (LLMs) and their emergent behaviors. He challenges the hype surrounding these models' capabilities, arguing that nonlinear metrics can create illusions of rapid progress. The conversation also discusses his work on trustworthiness in AI, focusing on critical aspects like toxicity and fairness. Sanmi highlights the need for robust evaluation methods as LLMs are integrated into sensitive fields like healthcare and education.

46 snips
Feb 5, 2024 • 1h 10min
AI Trends 2024: Reinforcement Learning in the Age of LLMs with Kamyar Azizzadenesheli - #670
Kamyar Azizzadenesheli, a staff researcher at Nvidia specializing in reinforcement learning, shares exciting insights on the collaboration between RL and large language models. He discusses innovations like ALOHA, a robot learning to fold clothes, and Voyager, an RL agent excelling in Minecraft using GPT-4. The conversation highlights advancements in risk-aware RL, especially in healthcare and finance. Kamyar also predicts how enhanced computational power will shape the future of deep reinforcement learning and facilitate general intelligence.

65 snips
Jan 29, 2024 • 35min
Building and Deploying Real-World RAG Applications with Ram Sriharsha - #669
Ram Sriharsha, VP of Engineering at Pinecone and an expert in large-scale data processing, explores the transformative power of vector databases and retrieval augmented generation (RAG). He discusses the trade-offs between LLMs and vector databases for effective data retrieval. The conversation sheds light on the evolution of RAG applications, the complexities of maintaining fresh enterprise data, and the exciting new features of Pinecone's serverless offering, which enhances scalability and cost efficiency. Ram also shares insights on the future of vector databases in AI.

6 snips
Jan 22, 2024 • 40min
Nightshade: Data Poisoning to Fight Generative AI with Ben Zhao - #668
In this engaging conversation, Ben Zhao, a Neubauer professor of computer science at the University of Chicago, dives into the critical intersection of security and generative AI. He introduces innovative tools like Fawkes, which masks images from facial recognition, and Glaze, designed to protect artists from style mimicry by subtly altering their work. Zhao also unveils Nightshade, a sophisticated defense mechanism that disrupts generative AI's ability to replicate artistic creations, raising vital questions about data poisoning and copyright in the AI era.
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.