
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.
Latest episodes

11 snips
Jun 3, 2024 • 48min
Energy Star Ratings for AI Models with Sasha Luccioni - #687
Sasha Luccioni, AI and Climate lead at Hugging Face, dives into the environmental impact of AI models. She discusses her groundbreaking research on energy consumption, revealing stark contrasts between generative and task-specific models. The conversation highlights the importance of a standardized Energy Star rating system for AI models, aiming to guide users towards energy-efficient choices. Luccioni also tackles challenges in evaluating model performance and the need for transparency and ethical standards in AI research to promote sustainable practices.

122 snips
May 27, 2024 • 56min
Language Understanding and LLMs with Christopher Manning - #686
Christopher Manning, a leading figure in machine learning and NLP from Stanford University, dives into the fascinating world of language models. He discusses the balance between linguistics and machine learning, emphasizing how LLMs learn human language structures. The talk covers the evolution and impact of word embeddings and attention mechanisms, along with the reasoning capabilities of these models. Manning also shares insights on emerging architectures and the future of AI research, making for an enlightening conversation on language understanding.

66 snips
May 20, 2024 • 43min
Chronos: Learning the Language of Time Series with Abdul Fatir Ansari - #685
In this discussion, machine learning scientist Abdul Fatir Ansari from AWS AI Labs dives into his groundbreaking work, Chronos, which applies language models to time series forecasting. He reveals the competitive edge Chronos has over traditional statistical methods and its surprising success in zero-shot forecasting. The conversation also touches on practical challenges like data augmentation and evaluation setups, as well as ongoing efforts to enhance synthetic data quality. Ansari sheds light on the promising future for integrating Chronos into real-world applications.

May 13, 2024 • 55min
Powering AI with the World's Largest Computer Chip with Joel Hestness - #684
In this discussion, Joel Hestness, a principal research scientist and lead of the core machine learning team at Cerebras, dives into the groundbreaking Wafer Scale Engine 3. He explains how this custom silicon surpasses traditional AI hardware, focusing on its unique architecture and memory capabilities. Joel also covers advancements in large language model training, innovative optimization techniques, and the integration of open-source frameworks like PyTorch. Additionally, he shares exciting research on weight-sparse training and novel optimizers that leverage higher-order statistics.

24 snips
May 7, 2024 • 50min
AI for Power & Energy with Laurent Boinot - #683
Laurent Boinot, Power and Utilities Lead for the Americas at Microsoft, dives into the fascinating intersection of AI and energy infrastructure. He discusses the challenges of North America's power systems and how AI is enhancing demand forecasting and grid optimization. Highlights include AI's role in securing utility systems and improving customer interactions. Laurent also explores the future of nuclear power and the pivotal role of electric vehicles in American energy management. It's a glimpse into a smarter, more efficient energy future!

Apr 29, 2024 • 42min
Controlling Fusion Reactor Instability with Deep Reinforcement Learning with Aza Jalalvand - #682
Aza Jalalvand, a research scholar at Princeton University, dives into the fascinating realm of using deep reinforcement learning to stabilize plasma in nuclear fusion reactors. He discusses the development of a model to combat the tearing mode instability while collecting complex data from fusion experiments. Aza highlights the critical role of machine learning in enhancing plasma understanding, the challenges of real-time data management, and the promising future of AI in clean energy production. Tune in for insights on the electrifying intersection of AI and fusion technology!

83 snips
Apr 22, 2024 • 47min
GraphRAG: Knowledge Graphs for AI Applications with Kirk Marple - #681
In this discussion, Kirk Marple, CEO of Graphlit, dives into the innovative world of GraphRAG, which combines knowledge graphs with retrieval-augmented generation. He shares insights on building knowledge graphs and the challenges of entity extraction. The conversation covers dynamic prompting techniques that enhance AI model responses and the integration of multiple storage types for effective data management. Kirk also explores the future of AI agents and their applications, showcasing the evolution of cloud services in graph-based technologies.

16 snips
Apr 16, 2024 • 46min
Teaching Large Language Models to Reason with Reinforcement Learning with Alex Havrilla - #680
In this engaging discussion, Alex Havrilla, a PhD student at Georgia Tech, dives into his research on enhancing reasoning in large language models using reinforcement learning. He explains the importance of creativity and exploration in AI problem-solving. Alex also highlights his findings on the effects of noise during training, revealing how resilient models can be. The conversation touches on the potential of combining language models with traditional methods to bolster AI reasoning, offering a glimpse into the exciting future of reinforcement learning.

79 snips
Apr 8, 2024 • 50min
Localizing and Editing Knowledge in LLMs with Peter Hase - #679
Peter Hase, a fifth-year PhD student at the University of North Carolina NLP lab, dives into the fascinating world of large language models. He discusses the vital role of interpretability in AI, exploring how knowledge is stored and accessed. The conversation shifts to model editing, emphasizing the challenges of deleting sensitive information while maintaining data integrity. Hase also highlights the risks of easy-to-hard generalization in releasing open-source models and the impact of instructional prompts on model performance. This insightful dialogue unravels complexities in AI decision-making.

103 snips
Apr 1, 2024 • 48min
Coercing LLMs to Do and Reveal (Almost) Anything with Jonas Geiping - #678
Jonas Geiping, a research group leader at the ELLIS Institute and Max Planck Institute, sheds light on his groundbreaking work on coercing large language models (LLMs). He discusses the alarming potential for LLMs to engage in harmful actions when misused. The conversation dives into the evolving landscape of AI security, exploring adversarial attacks and the significance of open models for research. They also touch on the complexities of input optimization and the balance between safeguarding models while maintaining their functionality.
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.