The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Latest episodes

undefined
Apr 16, 2024 • 46min

Teaching Large Language Models to Reason with Reinforcement Learning with Alex Havrilla - #680

PhD student Alex Havrilla from Georgia Tech talks about using reinforcement learning to improve reasoning in large language models. He discusses the role of creativity in problem solving, applying RL algorithms to enhance reasoning, noise's effect on language model training, and potential future developments in AI reasoning.
undefined
Apr 8, 2024 • 50min

Localizing and Editing Knowledge in LLMs with Peter Hase - #679

Peter Hase, a PhD student, discusses scalable oversight in neural networks, knowledge localization in LLMs, and the importance of deleting sensitive information. They explore interpretability techniques, surgical model editing, and task specification in pre-trained models, highlighting challenges in updating model knowledge and defending against information extraction.
undefined
Apr 1, 2024 • 48min

Coercing LLMs to Do and Reveal (Almost) Anything with Jonas Geiping - #678

Jonas Geiping, a research group leader at the ELLIS Institute, discusses the risks of deploying LLM agents, challenges in optimizing constraints, and the future of AI security. They explore vulnerabilities in LLMs, optimal text sequence generation, hybrid optimization strategies, reinforcement learning impact on model vulnerability, and enhancing safety through scaling models to prevent exploitative attacks.
undefined
Mar 25, 2024 • 48min

V-JEPA, AI Reasoning from a Non-Generative Architecture with Mido Assran - #677

The podcast delves into V-JEPA, a model for artificial reasoning bridging human-machine intelligence gap. It uses self-supervised training from unlabeled video data to learn abstract concepts efficiently. The discussion explores the development process, potential revolution in AI, and the shift towards predictive rather than generative models.
undefined
Mar 18, 2024 • 50min

Video as a Universal Interface for AI Reasoning with Sherry Yang - #676

Sherry Yang discusses the potential of video models in AI reasoning, comparing them to language models. They explore challenges in using video data for AI reasoning, integration of language and video models, and implications of video generation in scientific domains. The conversation also touches on the evolution of video models, interactive simulation, and manipulation in AI reasoning.
undefined
Mar 11, 2024 • 40min

Assessing the Risks of Open AI Models with Sayash Kapoor - #675

Discussing the risks and benefits of open AI models, including biosecurity threats and non-consensual imagery. Exploring a risk assessment framework inspired by cybersecurity. Emphasizing the need for common ground in assessing threats posed by AI. Addressing the balance between openness for research and cybersecurity vulnerabilities.
undefined
Mar 4, 2024 • 32min

OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia - #674

Akshita Bhagia discusses the OLMo language model with a unique open-source approach. The OLMo umbrella includes projects like Dolma and Paloma. The importance of open-training datasets and data curation filters are emphasized. The podcast explores dataset contamination, task specificity, and the evolution of training data transparency.
undefined
Feb 26, 2024 • 25min

Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673

A Stanford PhD student discusses reasoning in language models, emphasizing the importance of training data and chain-of-thought reasoning. The conversation explores human reasoning processes and the impact of prompts on enhancing language models through data curation.
undefined
Feb 19, 2024 • 46min

Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672

Armineh Nourbakhsh from JP Morgan AI Research discusses the development of DocLLM, a layout-aware large language model for document understanding. Topics include challenges of document AI, training approaches, datasets used, incorporating layout information, and evaluating model performance.
undefined
Feb 12, 2024 • 1h 6min

Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671

Sanmi Koyejo, assistant professor at Stanford University, discusses his award-winning papers on emergent abilities of large language models (LLMs) and assessing trustworthiness in GPT models. We explore the illusion of LLMs' rapid improvement and the importance of linear metrics. The methodology for evaluating concerns like toxicity and fairness in LLMs is also discussed. Personalized evaluation tests, tracking cross-metrics, and evaluating black box models are additional topics covered.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode