2min chapter

MLOps.community  cover image

Python Power: How Daft Embeds Models and Revolutionizes Data Processing // Sammy Sidhu // MLOps Podcast #165

MLOps.community

CHAPTER

How to Distilate a Large Language Model

There's this great paper from Alex Radner and a lot of other people that I can't remember who, but he was talking about distilled step by step. It's basically distilling the model. And so it makes it much easier for you to get that distilled model and train it with less data. Back in my day, distillation was kind of dumb. What you would do is you would train a big model, and you would just train the essentially, you would use that output as the ground truth for a smaller model. But now they're asking for this chain of thought reasoning.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode