The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

The Evolution of Reasoning in Small Language Models with Yejin Choi - #761

74 snips
Jan 29, 2026
Yejin Choi, Stanford professor focused on reasoning, synthetic data, and AI alignment, discusses making small language models reason better. She covers synthetic data generation, imitation learning, reinforcement pretraining that encourages internal “thinking,” and Prismatic Synthesis for diverse math data. The conversation also tackles mode collapse, spectrum tuning to preserve diversity, and pluralistic alignment to reflect varied human values.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Democratize Reasoning In Small Models

  • Yejin Choi focuses on making small language models (SLMs) reason better and democratizing generative AI beyond big companies.
  • She believes smaller models can gain significant capabilities with targeted research and data effort.
INSIGHT

Data As A Shortcut Over Scale

  • Many routes exist to make small models stronger: compression, new architectures, and better data.
  • High-quality, outskirt data often teaches small models reasoning faster than scale alone.
ADVICE

Engineer Synthetic Data Pipelines

  • Don't rely on vanilla synthetic generation; design prompts and pipelines to revise and improve outputs.
  • Use iterative prompting, model chains, and expert curation to create qualitatively new training data.
Get the Snipd Podcast app to discover more snips from this episode
Get the app