Machine Learning Street Talk (MLST)

The Lottery Ticket Hypothesis with Jonathan Frankle

May 19, 2020
Jonathan Frankle, author of The Lottery Ticket Hypothesis, shares his insights on Sparse Neural Networks and their pruning techniques. He delves into the implications of the lottery ticket hypothesis for improving neural network efficiency and discusses innovative strategies like linear mode connectivity. Frankle also explores the intersection of AI technology and policy, emphasizing the importance of ethical decision-making in AI development. Listeners will appreciate his journey in deep learning research and the challenges faced in academia.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Smaller Networks After Training

  • Neural networks can be made smaller after training without significant accuracy loss.
  • This suggests the learned function has a smaller representation than the initial network.
INSIGHT

Instability and Lottery Ticket Behavior

  • Instability in training, caused by SGD noise, affects lottery ticket behavior.
  • If a subnetwork consistently trains to the same optimum, it performs well; inconsistency leads to worse performance.
ANECDOTE

Learning Rate Rewinding

  • In pruning research, resetting learning rates without resetting weights after pruning improves performance.
  • This learning rate rewinding approach matches state-of-the-art pruning techniques.
Get the Snipd Podcast app to discover more snips from this episode
Get the app