
Attention Primer
Data Skeptic
00:00
Is Attention Important for Machine Learning?
There's an idea that you can only hold so many things in your memory at once. How do we make machines that not only have this ability, to take details from earlier, but also learn the model or the mechanism for which things are actually important? So maybe it would be helpful if i described l bit more about how machine learning is sorof the plain vanilla versions are kind of limited.
Transcript
Play full episode