Gradient Dissent: Conversations on AI cover image

Neural Network Pruning and Training with Jonathan Frankle at MosaicML

Gradient Dissent: Conversations on AI

00:00

The Effects of Dropout on the Network

The claim that it's necessarily leaning on that 10%, I think is something that we can conjecture about. It's hard to say which weights are actually important and which weights aren't for the function. Using magnitude as a heuristic is a very bad one, at least. There does tend to be a huge difference between pruning weights and pruning neurons in terms of how well you do with them.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app