MLOps.community

Efficient GPU infrastructure at LinkedIn // Animesh Singh // MLOps Podcast #299

Mar 28, 2025
Animesh Singh, Executive Director of AI and ML Platform at LinkedIn, leads the charge in evolving AI technologies. He dives into the transformative impact of large language models on recruitment, highlighting LinkedIn's Hiring Assistant. Animesh also discusses the financial challenges of GPU infrastructure, emphasizing the need for optimization strategies. The conversation touches on real-time training and the intricate balance between scaling AI advancements and managing costs, offering insights into the future of AI and infrastructure innovations.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

LLMs at LinkedIn

  • Animesh Singh discusses how the launch of ChatGPT changed LinkedIn's AI focus.
  • He describes new LLM-powered features like profile summarization and personalized recruiter emails.
INSIGHT

LLM Cost Bottleneck

  • Inferencing costs, not training costs, are a major hurdle for LLM adoption.
  • Open-source models and fine-tuning have reduced training costs, but inferencing remains expensive.
INSIGHT

GPU Underutilization

  • Despite infrastructure investments, GPUs underperform in inferencing due to latency and throughput optimization.
  • Real-time applications like recommendation systems have no latency tolerance, making LLMs costly.
Get the Snipd Podcast app to discover more snips from this episode
Get the app