Practical AI

Optimizing for efficiency with IBM’s Granite

167 snips
Mar 14, 2025
Kate Soule, Director of Technical Product Management at IBM's Granite, shares insights on the evolution of AI models. She emphasizes the importance of efficiency over conventional leaderboard scores. The conversation explores breaking tasks into manageable components, co-designing models with hardware, and the innovative 'mixture of experts' approach for speed and efficiency. Soule also discusses the shift towards open-source models, ethical considerations in AI, and the exciting advancements in Granite 3.2, marking a new frontier in performance.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Kate Soule's Background

  • Kate Soule leads product management for Granite, IBM's large language models.
  • She joined IBM Research years before LLMs became popular, coming from a business background.
INSIGHT

LLMs as Building Blocks

  • LLMs are reusable building blocks for various applications, reducing the need for bespoke models.
  • IBM centralizes LLM development in research to support diverse use cases and open-source collaboration.
INSIGHT

Why Open Source?

  • IBM chose Apache 2 license for Granite to promote broad usage and customization.
  • Open source fosters stronger software and aligns with IBM's strategy, building upon their experience with Red Hat.
Get the Snipd Podcast app to discover more snips from this episode
Get the app