The Gradient: Perspectives on AI cover image

Hattie Zhou: Lottery Tickets and Algorithmic Reasoning in LLMs

The Gradient: Perspectives on AI

00:00

How Do You Set Weights to Zero?

I'm curious about when it comes to the setting weights to zero is what the picture might look like for really large networks or larger networks than you were looking at. One thing I can imagine in line with the idea that perhaps a model over its training evolution perhaps forgets things, but then it also relearns them kind of not so subtly making reference to another paper we're going to talk about soon. At some point in training, maybe a set of weights start converging to zero, but then later in its training dynamics, it realizes oh wait, these were actually important for something or the other.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app