Interconnects

Elicitation, the simplest way to understand post-training

12 snips
Mar 10, 2025
Discover how the concept of elicitation can dramatically enhance AI model performance after training. The discussion uses a thrilling Formula 1 analogy to illustrate how teams optimize their cars throughout a season, showing similar potential in AI models. The conversation also touches on the Superficial Alignment Hypothesis, emphasizing the importance of pre-existing data. Join in to explore innovative techniques that can lead to significant improvements in a short time frame!
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Post-Training Improvements

  • Many recent model improvements from major AI companies have been in post-training.
  • This phase extracts and amplifies valuable behaviors already present in the base model.
ANECDOTE

F1 Analogy

  • F1 teams make significant performance gains through aerodynamic and system changes throughout the season, similar to post-training in AI.
  • OLMoE Instruct improved from 35 to 48 in post-training evaluation without changing the majority of pre-training.
INSIGHT

Scaling and Post-Training

  • Larger models, like GPT-4.5, offer a more dynamic base for post-training improvements.
  • Scaling allows for faster post-training but requires significant infrastructure investments.
Get the Snipd Podcast app to discover more snips from this episode
Get the app