Super Data Science: ML & AI Podcast with Jon Krohn cover image

747: Technical Intro to Transformers and LLMs, with Kirill Eremenko

Super Data Science: ML & AI Podcast with Jon Krohn

00:00

Technical Aspects of Transformers and LLMs

Exploration of the technical components of Transformers and Large Language Models (LLMs), highlighting incremental changes in attention heads, context windows, and parameter settings since the publication of the key paper 'Attention Is All You Need' in 2017. The chapter provides an overview of the encoder and decoder parts in transformers, explaining their specific uses in models like GPT and BERT depending on the task at hand.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app