Backpropagation
Linear Digressions
00:00
Neural Nets and Backpropagation
Backpropagation is the way of taking mistakes that you make and learning from them, basically. So there's an emergent property in a way where the basic operating principles are relatively simple. But when you put it all together, you get something a bit more complex from stacking these things. And so making those adjustments is how you actually train the neural net.
Play episode from 06:16
Transcript


