Gradient Dissent: Conversations on AI cover image

Neural Network Pruning and Training with Jonathan Frankle at MosaicML

Gradient Dissent: Conversations on AI

00:00

The Lottery Ticket Hypothesis

Are you saying that the point of having many more weights than you need is just that some of them randomly get assigned good initialization values? Is that what you're saying? It's a possibility. The only reason I'll say it's a possibility is because I'm an empiricist. If I'm going to make a claim, I need to have an experiment to evaluate it and try to falsify it.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app