Unsupervised Learning cover image

Unsupervised Learning

Ep 64: GPT 4.1 Lead at OpenAI Michelle Pokrass: RFT Launch, How OpenAI Improves Its Models & the State of AI Agents Today

May 8, 2025
Michelle Pokrass, a key player in launching GPT-4.1 at OpenAI, delves into the intricacies of AI model development and evaluation. She discusses how user feedback fuels improvements and the evolving challenges in AI training. The conversation highlights the shift towards Reinforcement Fine-Tuning for personalized applications and the importance of balancing conversational abilities with problem-solving as we look towards future models. Her insights reflect a deep understanding of AI's potential and the organizational growth at OpenAI.
47:12

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • The development of GPT-4.1 emphasizes real-world usability through direct user feedback, shaping evaluation methods and training strategies.
  • Fine-tuning strategies, particularly Reinforcement Fine-Tuning (RFT), enhance model performance by improving data efficiency and tailoring solutions.

Deep dives

Enhancements in GPT 4.1 for Developer Utility

GPT 4.1 places a stronger emphasis on real-world usage rather than merely excelling in benchmarks. Developers sought improvements in areas such as instruction adherence and formatting consistency. To achieve this, the development team directly engaged with users to gather feedback, which informed their evaluation methods before model training began. The result is an improved model that better aligns with developers' practical needs, ultimately delivering a more satisfying user experience.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app