
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.
Latest episodes

143 snips
Jul 23, 2024 • 1h 20min
Building Real-World LLM Products with Fine-Tuning and More with Hamel Husain - #694
In this discussion with Hamel Husain, founder of Parlance Labs, they dive into the practicalities of leveraging large language models (LLMs) for real-world applications. Husain shares insights on fine-tuning techniques, including tools like Axolotl and the advantages of LoRa for efficient model adjustments. He emphasizes the importance of thoughtful user interface design and systematic evaluation strategies to enhance AI's effectiveness. The conversation also highlights challenges in data curation and the need for accurate metrics in domain-specific projects, ensuring robust AI development.

20 snips
Jul 17, 2024 • 58min
Mamba, Mamba-2 and Post-Transformer Architectures for Generative AI with Albert Gu - #693
In this discussion, Albert Gu, an assistant professor at Carnegie Mellon University, dives into his research on post-transformer architectures. He explains the efficiency and challenges of the attention mechanism, particularly in managing high-resolution data. The conversation highlights the significance of tokenization in enhancing model effectiveness. Gu also explores hybrid models that blend attention with state-space elements and emphasizes the groundbreaking advancements brought by his Mamba and Mamba-2 frameworks. His vision for the future of multi-modal foundation models is both insightful and inspiring.

Jul 9, 2024 • 43min
Decoding Animal Behavior to Train Robots with EgoPet with Amir Bar - #692
Join Amir Bar, a PhD candidate at Tel Aviv University and UC Berkeley, as he unpacks his groundbreaking research on visual-based learning and self-supervised object detection. He introduces ‘EgoPet,’ a unique dataset that captures animal behavior from their perspective, aiming to bridge the gap between AI and nature. The discussion dives into challenges of current classification methods, the significance of ego-centric data in robotic training, and the potential to enhance robotic navigation by mimicking animal locomotion. Exploration of these topics reveals fascinating insights into future AI advancements.

9 snips
Jul 1, 2024 • 57min
How Microsoft Scales Testing and Safety for Generative AI with Sarah Bird - #691
Join Sarah Bird, Chief Product Officer of Responsible AI at Microsoft, as she dives into the essential realms of generative AI testing and safety. Explore the challenges of AI hallucinations and the importance of balancing fairness with security. Hear insights from Microsoft's past failures like Tay and Bing Chat, stressing the need for adaptive testing and human oversight. Sarah also discusses innovative methods like automated safety testing and red teaming, emphasizing a robust governance framework for evolving AI technologies.

9 snips
Jun 25, 2024 • 46min
Long Context Language Models and their Biological Applications with Eric Nguyen - #690
Eric Nguyen, a PhD student at Stanford, dives deep into his research on long context foundation models, specifically Hyena and its applications in biology. He explains the limitations of traditional transformers in processing lengthy sequences and how convolutional models provide innovative solutions. Nguyen introduces Hyena DNA, designed for long-range DNA dependencies, and discusses Evo, a hybrid model with massive parameters for DNA generation. The podcast touches on exciting applications in CRISPR gene editing and the implications of using AI in biological research.

Jun 18, 2024 • 48min
Accelerating Sustainability with AI with Andres Ravinet - #689
In this engaging discussion, Andres Ravinet, Sustainability Global Black Belt at Microsoft, shares his insights on harnessing AI for sustainability challenges. He highlights innovative AI applications, like early warning systems for extreme weather and methods to reduce food waste in supply chains. Ravinet also addresses the complexities of ESG compliance reporting and the driving forces behind corporate sustainability efforts. Additionally, he explores how generative AI can further support these initiatives, showcasing a commitment to a greener future.

Jun 10, 2024 • 1h 11min
Gen AI at the Edge: Qualcomm AI Research at CVPR 2024 with Fatih Porikli - #688
Fatih Porikli, Senior Director of Technology at Qualcomm AI Research, dives into groundbreaking advancements in generative AI and computer vision. He discusses efficient diffusion models for text-to-image generation and real-time 360° image relighting. The conversation also highlights innovative applications like a video-language model for personalized fitness coaching and a Math Search dataset for visual reasoning. Porikli touches on practical demos at CVPR, showcasing multi-modal models and enhancing AI's capabilities for mobile and edge devices.

11 snips
Jun 3, 2024 • 48min
Energy Star Ratings for AI Models with Sasha Luccioni - #687
Sasha Luccioni, AI and Climate lead at Hugging Face, dives into the environmental impact of AI models. She discusses her groundbreaking research on energy consumption, revealing stark contrasts between generative and task-specific models. The conversation highlights the importance of a standardized Energy Star rating system for AI models, aiming to guide users towards energy-efficient choices. Luccioni also tackles challenges in evaluating model performance and the need for transparency and ethical standards in AI research to promote sustainable practices.

122 snips
May 27, 2024 • 56min
Language Understanding and LLMs with Christopher Manning - #686
Christopher Manning, a leading figure in machine learning and NLP from Stanford University, dives into the fascinating world of language models. He discusses the balance between linguistics and machine learning, emphasizing how LLMs learn human language structures. The talk covers the evolution and impact of word embeddings and attention mechanisms, along with the reasoning capabilities of these models. Manning also shares insights on emerging architectures and the future of AI research, making for an enlightening conversation on language understanding.

76 snips
May 20, 2024 • 43min
Chronos: Learning the Language of Time Series with Abdul Fatir Ansari - #685
In this discussion, machine learning scientist Abdul Fatir Ansari from AWS AI Labs dives into his groundbreaking work, Chronos, which applies language models to time series forecasting. He reveals the competitive edge Chronos has over traditional statistical methods and its surprising success in zero-shot forecasting. The conversation also touches on practical challenges like data augmentation and evaluation setups, as well as ongoing efforts to enhance synthetic data quality. Ansari sheds light on the promising future for integrating Chronos into real-world applications.