The Gradient: Perspectives on AI cover image

Sewon Min: The Science of Natural Language

The Gradient: Perspectives on AI

00:00

The Meta-Training of Language Models

It seems like model saturation is something that we tend to see a lot. And so it's interesting, but I guess not entirely surprising. The meta trained model ignores the input label mapping more than regular language models. It just helps the model to better locate the task that is already learned during training. But it perhaps seems to teach the model entirely new ability.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app