Latent Space: The AI Engineer Podcast cover image

Everything you need to run Mission Critical Inference (ft. DeepSeek v3 + SGLang)

Latent Space: The AI Engineer Podcast

00:00

Unpacking DeepSeek V3: Challenges and Innovations

This chapter explores the advancements of DeepSeek V3, a leading open-source language model with 671 billion parameters, and the hurdles associated with its deployment. The speakers delve into loading times, performance benchmarks, and user motivations for transitioning to in-house solutions. They also discuss the significance of FP8 quantization in model training, highlighting the evolving landscape of AI inference technologies.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app