
The Information Bottleneck EP13: Recurrent-Depth Models and Latent Reasoning with Jonas Geiping
Nov 7, 2025
Jonas Geiping, a machine learning researcher at the ELLIS Institute and Max Planck Institute, explores the fascinating world of recurrent-depth models and latent reasoning. He discusses how these models can enhance AI's reasoning capabilities, especially in complex tasks like math and coding. The conversation also delves into challenges in model development, the importance of interpretability and safety in AI, and the future of scalable algorithms. With practical advice for budding researchers, Jonas sheds light on Tübingen as an emerging hub for machine learning innovation.
AI Snips
Chapters
Transcript
Episode notes
One Month Lost To Prompting Claude
- Jonas spent a month trying to get Claude to implement a sampling method and eventually coded it himself.
- Writing the implementation personally was "healing" and faster than iterating with the model.
Use The Small-Scale Training Playbook
- Follow detailed training playbooks (data cleaning, mixing, optimizer, architecture choices) before large runs.
- Use small-scale experiments to discover pitfalls before burning major compute.
Adaptive Compute Through Recurrence
- Recurrent-depth models let models reuse the same layers multiple times to adapt compute per token.
- This enables deeper on-demand computation for hard tokens without a fixed large network per token.

