
747: Technical Intro to Transformers and LLMs, with Kirill Eremenko
Super Data Science: ML & AI Podcast with Jon Krohn
00:00
Technical Aspects of Transformers and LLMs
Exploration of the technical components of Transformers and Large Language Models (LLMs), highlighting incremental changes in attention heads, context windows, and parameter settings since the publication of the key paper 'Attention Is All You Need' in 2017. The chapter provides an overview of the encoder and decoder parts in transformers, explaining their specific uses in models like GPT and BERT depending on the task at hand.
Transcript
Play full episode