Collective Intelligence Community Podcast

OpenPipe Co-Founder on Reinforcement Learning - David Corbitt | Ep. 22

Jul 29, 2025
David Corbitt, co-founder and Chief Product Officer of OpenPipe, discusses innovative ways to enhance small AI models. He shares insights on fine-tuning and reinforcement learning that empower these models to outperform larger alternatives. The conversation reveals how OpenPipe reduces costs and latency, making AI more accessible. Listeners learn about the significance of realistic training scenarios and the role of human feedback in developing reliable AI. Corbitt also emphasizes the enduring value of specialized models in a landscape dominated by large, closed systems.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Small Models Beating Big Models

  • OpenPipe focuses on making small open models outperform large closed models for specific tasks.
  • This improves quality, cost, latency, and independence from central providers.
ADVICE

Fine-Tune When Tasks Diverge

  • Fine-tune models when your task differs from big labs' training data to increase reliability.
  • Switch to smaller fine-tuned models to reduce inference cost and latency.
INSIGHT

GPU Time Drives Costs

  • Training and inference costs are dominated by GPU time (H100/H200).
  • LoRA-style parameter-efficient tuning greatly reduced training costs.
Get the Snipd Podcast app to discover more snips from this episode
Get the app