Gradient Dissent: Conversations on AI cover image

Neural Network Pruning and Training with Jonathan Frankle at MosaicML

Gradient Dissent: Conversations on AI

00:00

The Power of Weights in Neural Networks

The story of my career in machine learning is questioning those choices. In the lottery ticket work, I questioned one very specific choice. Why do we use all these weights? These networks are really big. We know they're overparameterized, but why? The trick is that any weight you were going to delete at the end of training, you never really need it.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app