The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Latest episodes

undefined
29 snips
Mar 4, 2024 • 32min

OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia - #674

Akshita Bhagia, a senior research engineer at the Allen Institute for AI, shares her insights on OLMo, an open-source language model that includes a unique dataset and tools for training. She discusses the innovative Dolma dataset, which boasts a three-trillion-token corpus, and Paloma, a benchmarking tool for evaluating model performance. Throughout the conversation, Akshita emphasizes the importance of data transparency, collaborative research, and the challenges faced in training large-scale models, advocating for a shared knowledge approach in AI development.
undefined
6 snips
Feb 26, 2024 • 25min

Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673

Ben Prystawski, a PhD student at Stanford blending cognitive science with machine learning, unveils fascinating insights on LLM reasoning. He discusses his recent paper that questions if reasoning exists in LLMs and the effectiveness of chain-of-thought strategies. Delve into how locality in training data fuels reasoning capabilities, and explore the nuances of optimizing prompts for better model performance. The conversation also touches on how our human experiences shape reasoning, enhancing comprehension in artificial intelligence.
undefined
24 snips
Feb 19, 2024 • 46min

Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672

Armineh Nourbakhsh, Executive Director at JP Morgan AI Research, dives into the exciting world of DocLLM, a layout-aware large language model designed for document understanding. She shares insights on the evolution of document AI, focusing on multimodal approaches that combine textual and visual data. Nourbakhsh discusses the challenges of training generative models, the intricacies of processing enterprise documents, and strategies to reduce hallucinations in language models, enhancing performance in complex document analysis.
undefined
23 snips
Feb 12, 2024 • 1h 6min

Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671

Sanmi Koyejo, an assistant professor at Stanford University, dives into the fascinating world of large language models (LLMs) and their emergent behaviors. He challenges the hype surrounding these models' capabilities, arguing that nonlinear metrics can create illusions of rapid progress. The conversation also discusses his work on trustworthiness in AI, focusing on critical aspects like toxicity and fairness. Sanmi highlights the need for robust evaluation methods as LLMs are integrated into sensitive fields like healthcare and education.
undefined
46 snips
Feb 5, 2024 • 1h 10min

AI Trends 2024: Reinforcement Learning in the Age of LLMs with Kamyar Azizzadenesheli - #670

Kamyar Azizzadenesheli, a staff researcher at Nvidia specializing in reinforcement learning, shares exciting insights on the collaboration between RL and large language models. He discusses innovations like ALOHA, a robot learning to fold clothes, and Voyager, an RL agent excelling in Minecraft using GPT-4. The conversation highlights advancements in risk-aware RL, especially in healthcare and finance. Kamyar also predicts how enhanced computational power will shape the future of deep reinforcement learning and facilitate general intelligence.
undefined
65 snips
Jan 29, 2024 • 35min

Building and Deploying Real-World RAG Applications with Ram Sriharsha - #669

Ram Sriharsha, VP of Engineering at Pinecone and an expert in large-scale data processing, explores the transformative power of vector databases and retrieval augmented generation (RAG). He discusses the trade-offs between LLMs and vector databases for effective data retrieval. The conversation sheds light on the evolution of RAG applications, the complexities of maintaining fresh enterprise data, and the exciting new features of Pinecone's serverless offering, which enhances scalability and cost efficiency. Ram also shares insights on the future of vector databases in AI.
undefined
6 snips
Jan 22, 2024 • 40min

Nightshade: Data Poisoning to Fight Generative AI with Ben Zhao - #668

In this engaging conversation, Ben Zhao, a Neubauer professor of computer science at the University of Chicago, dives into the critical intersection of security and generative AI. He introduces innovative tools like Fawkes, which masks images from facial recognition, and Glaze, designed to protect artists from style mimicry by subtly altering their work. Zhao also unveils Nightshade, a sophisticated defense mechanism that disrupts generative AI's ability to replicate artistic creations, raising vital questions about data poisoning and copyright in the AI era.
undefined
Jan 15, 2024 • 39min

Learning Transformer Programs with Dan Friedman - #667

Dan Friedman, a PhD student from Princeton's NLP group, dives into his fascinating research on mechanistic interpretability for transformer models. He discusses his innovative paper that modifies transformer architecture to create human-readable programs. The conversation uncovers the challenges of current interpretability methods and contrasts them with his approach. They explore the RASP framework's role in transforming programs and delve into the complexities of optimizing model constraints, highlighting the importance of clarity in understanding AI.
undefined
77 snips
Jan 8, 2024 • 1h 5min

AI Trends 2024: Machine Learning & Deep Learning with Thomas Dietterich - #666

Thomas Dietterich, a distinguished professor emeritus at Oregon State University, dives into the latest trends in AI and machine learning. He discusses the strengths and weaknesses of large language models like GPT-4, while exploring their potential limitations in reasoning. The conversation covers topics like uncertainty quantification and the fascinating world of 'hallucinations' in language models. Dietterich also offers predictions for 2024 and motivates newcomers to tap into the field's endless possibilities.
undefined
15 snips
Jan 2, 2024 • 52min

AI Trends 2024: Computer Vision with Naila Murray - #665

Naila Murray, Director of AI Research at Meta, discusses the cutting-edge landscape of computer vision. They explore advancements like controllable AI generation, multimodal models, and tools such as Segment Anything for intuitive image segmentation. Naila dives into the possibilities of ControlNet and DINOv2, emphasizing their roles in object recognition and complex scenarios. Looking ahead, she shares insights on opportunities in self-supervised learning and generative models, forecasting exciting innovations for 2024 in AI.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app