
 The MAD Podcast with Matt Turck
 The MAD Podcast with Matt Turck How GPT-5 Thinks — OpenAI VP of Research Jerry Tworek
 278 snips 
 Oct 16, 2025  Join Jerry Tworek, VP of Research at OpenAI, as he dives into the fascinating world of AI reasoning. Discover how GPT-5 evolves from earlier models, emphasizing the crucial roles of pretraining and reinforcement learning. Jerry explains the mechanics of chain-of-thought reasoning, the significance of agentic tools like Codex, and the importance of robust collaboration in research. He even shares insights from his personal journey from math and trading to cutting-edge AI research. Could pretraining combined with RL be key to achieving AGI? Tune in to find out! 
 AI Snips 
 Chapters 
 Transcript 
 Episode notes 
What AI Reasoning Actually Is
- Reasoning is the process of getting to answers you don't already know by performing intermediate work and search-like steps.
- Chain-of-thought verbalizes that process in words, letting models show systematic intermediate steps to reach solutions.
How Chain-Of-Thought Helps
- Chain-of-thought prompts coax models to generate stepwise tokens that perform partial computations and then combine them into an answer.
- Models learn these human-like stepwise procedures from pretraining on internet text containing many examples of human reasoning.
Optimize Thinking Time For UX
- Balance compute (thinking time) versus user wait time by exposing options for higher or lower 'thinking' settings to users.
- Tune heuristics to predict when extra thinking is worth the user's waiting time and offer auto modes accordingly.
