The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Latest episodes

undefined
Nov 19, 2024 • 54min

AI for Network Management with Shirley Wu - #710

Shirley Wu, Senior Director of Software Engineering at Juniper Networks, leads the discussion on the transformative power of AI and ML in network management. She shares insights on diagnosing cable issues and proactive coverage monitoring, highlighting the benefits of integrating data science. The conversation covers the evolution of roles for network administrators and the innovative Marvis chatbot, which enhances user engagement. Shirley also discusses future directions like proactive testing and the application of smaller ML models to improve efficiency and cost management.
undefined
Nov 11, 2024 • 58min

Why Your RAG System Is Broken, and How to Fix It with Jason Liu - #709

Jason Liu, a freelance AI consultant and creator of the Instructor library, dives deep into retrieval-augmented generation (RAG) systems. He shares key signs of RAG system failures and the tactical steps to diagnose issues. Liu emphasizes building robust test datasets and the importance of data-driven experimentation. He discusses fine-tuning strategies, chunking techniques, and collaboration tools, while also showcasing how future AI models could revolutionize the space. Finally, he highlights his passion for teaching through his AI consulting course.
undefined
Nov 4, 2024 • 1h 15min

An Agentic Mixture of Experts for DevOps with Sunil Mallya - #708

Today we're joined by Sunil Mallya, CTO and co-founder of Flip AI. We discuss Flip’s incident debugging system for DevOps, which was built using a custom mixture of experts (MoE) large language model (LLM) trained on a novel "CoMELT" observability dataset which combines traditional MELT data—metrics, events, logs, and traces—with code to efficiently identify root failure causes in complex software systems. We discuss the challenges of integrating time-series data with LLMs and their multi-decoder architecture designed for this purpose. Sunil describes their system's agent-based design, focusing on clear roles and boundaries to ensure reliability. We examine their "chaos gym," a reinforcement learning environment used for testing and improving the system's robustness. Finally, we discuss the practical considerations of deploying such a system at scale in diverse environments and much more.The complete show notes for this episode can be found at https://twimlai.com/go/708.
undefined
Oct 28, 2024 • 1h 2min

Building AI Voice Agents with Scott Stephenson - #707

Scott Stephenson, co-founder and CEO of Deepgram, shares his unique journey from particle physics to AI voice technology. He highlights the complexities of building intelligent voice agents, focusing on perception, interaction, and real-time updates. The discussion dives into the transformative potential of AI in customer service, emphasizing federated learning for continuous improvement. Scott also unveils Deepgram's new agent toolkit, showcasing applications across industries like healthcare and food service, and the need for adaptable models in voice interactions.
undefined
Oct 21, 2024 • 56min

Is Artificial Superintelligence Imminent? with Tim Rocktäschel - #706

Tim Rocktäschel, senior staff research scientist at Google DeepMind and AI professor at UCL, explores the tantalizing prospects of artificial superintelligence. He discusses the journey from narrow AI to superhuman capabilities, stressing the necessity of open-ended system development. The conversation also dives into the transformative impact of AI in science and medicine, alongside its role in enhancing debate automation for truth-seeking. With insights from his recent research, he highlights the importance of evolutionary algorithms and addresses challenges like bias in AI.
undefined
Oct 14, 2024 • 1h 16min

ML Models for Safety-Critical Systems with Lucas García - #705

Lucas García, Principal Product Manager for Deep Learning at MathWorks, dives into the integration of ML in safety-critical systems. He discusses crucial verification and validation processes, highlighting the V-model and its W-adaptation for ML. The conversation shifts to deep learning in aviation, focusing on data quality, model robustness, and interpretability. Lucas also introduces constrained deep learning and convex neural networks, examining the benefits and trade-offs of these approaches while stressing the importance of safety protocols and regulatory frameworks.
undefined
Oct 7, 2024 • 54min

AI Agents: Substance or Snake Oil with Arvind Narayanan - #704

Join Arvind Narayanan, a Princeton professor and expert on AI agents and policy, as he unpacks the substance behind AI technology. He discusses the risks of deploying AI agents and the pressing need for better benchmarking to ensure reliability. Delve into his book, which exposes exaggerated AI claims and failed applications. Narayanan also highlights his work on CORE-Bench, aiming to enhance scientific reproducibility and reviews the complex landscape of AI reasoning methods. He wraps up with insights on the tangled web of AI regulation and policy challenges.
undefined
Sep 30, 2024 • 48min

AI Agents for Data Analysis with Shreya Shankar - #703

Shreya Shankar, a PhD student at UC Berkeley, shares her insights on DocETL, a system designed for optimizing LLM-powered data processing pipelines. They discuss the fascinating challenges of intelligent data extraction from unstructured sources like PDFs and the pivotal role of human insight in prompt engineering. Shreya emphasizes the need for tailored benchmarks in data processing tasks and showcases real-world applications, including police misconduct data collection. The conversation highlights the balance between automation and human collaboration in AI systems.
undefined
Sep 23, 2024 • 1h 4min

Stealing Part of a Production Language Model with Nicholas Carlini - #702

Nicholas Carlini, a research scientist at Google DeepMind specializing in adversarial machine learning and model security, dives into model stealing techniques in this discussion. He reveals how parts of production language models like ChatGPT can be extracted, raising important ethical and security concerns. The episode highlights the current landscape of AI security and the steps tech giants are taking to protect against vulnerabilities. Carlini also shares insights from his best paper on privacy challenges in public pretraining and the complexities surrounding differential privacy.
undefined
Sep 16, 2024 • 1h 14min

Supercharging Developer Productivity with ChatGPT and Claude with Simon Willison - #701

In this conversation with Simon Willison, an independent researcher and creator of Datasette, we discover how developers can supercharge productivity using LLMs like ChatGPT and Claude. Simon shares his innovative workflows, including coding while walking his dog. He dives into effective prompting techniques, rapid prototyping with Claude’s Artifacts feature, and the evolving role of open-source and local models. The discussion also touches on the transformative impact of AI on coding practices and modern web scraping tools.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode