The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671

Feb 12, 2024
Sanmi Koyejo, an assistant professor at Stanford University, dives into the fascinating world of large language models (LLMs) and their emergent behaviors. He challenges the hype surrounding these models' capabilities, arguing that nonlinear metrics can create illusions of rapid progress. The conversation also discusses his work on trustworthiness in AI, focusing on critical aspects like toxicity and fairness. Sanmi highlights the need for robust evaluation methods as LLMs are integrated into sensitive fields like healthcare and education.
01:05:40

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • Linear metrics show smooth improvement in model performance, casting doubt on the significance of emergent abilities in large language models.
  • DecodingTrust methodology provides a comprehensive assessment of trustworthiness in GPT models, evaluating concerns like toxicity, privacy, fairness, and robustness.

Deep dives

Research Interests and Papers

Sammy Criagev, an assistant professor at Stanford University, discusses his research agenda focused on trustworthy AI systems. His lab explores foundational aspects, measurement and assessment, as well as mitigation strategies. The lab has recently delved into the study of language models and the emergent properties that arise as these models scale in size.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner