Super Data Science: ML & AI Podcast with Jon Krohn

759: Full Encoder-Decoder Transformers Fully Explained, with Kirill Eremenko

Feb 20, 2024
In this podcast, Kirill Eremenko, the SuperDataScience founder, discusses full encoder-decoder transformers with Jon Krohn. They cover topics like how cross-attention works, the importance of masking during self-attention, and the collaboration dynamics in transformer research. The episode provides a detailed explanation of encoder-decoder transformers, language models, and the use of transformers in natural language processing.
Ask episode
Chapters
Transcript
Episode notes