

Neural Network Pruning and Training with Jonathan Frankle at MosaicML
Apr 4, 2023
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Introduction
00:00 • 3min
The Power of Weights in Neural Networks
02:35 • 3min
The Lottery Ticket Hypothesis
05:36 • 2min
The Sub-Network and the Dense Network
07:24 • 2min
The Effects of Dropout on the Network
08:59 • 2min
How Much Pruning Can You Prune?
11:00 • 2min
The Unscientific Nature of Neural Networks
13:12 • 2min
Transformers and Attention: A Simple Architecture
15:31 • 2min
How to Speed Up Training With ResNet 50
17:50 • 4min
Mosaic and Mal: The Story Behind the Company
21:52 • 3min
How to Train a Machine Learning Model
25:06 • 2min
How to Engage With the Research Community
27:35 • 3min
The Skeptics of AGI
30:48 • 2min
The Importance of Feed Forward Networks
32:25 • 2min
The Future of GPT Chat
33:59 • 2min
The Importance of Policy
35:44 • 3min
The Future of Language and Vision Models
39:04 • 2min
The Importance of Adapting to Changes in the Chip World
41:03 • 3min
The Future of Diffusion Models
44:01 • 2min
The Importance of Data in Mosaic
45:56 • 4min
The Importance of Data in Large Language Model Training
49:41 • 2min
The Future of Data Curation and Data Labeling
51:20 • 2min
The Importance of Data Quality in Machine Learning
53:19 • 2min
The Unexpected Bottlenecks in Modeling
55:43 • 7min