

747: Technical Intro to Transformers and LLMs, with Kirill Eremenko
Jan 9, 2024
Data scientist Kirill Eremenko discusses the basics of transformers and LLMs, emphasizing the five building blocks of transformer architecture and why transformers are so powerful. Topics include AI recruitment, a new course on LLMs, and the impact of LLMs on data science jobs.
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Intro
00:00 • 2min
Transition of Hosting and Excitement for Milestone
02:18 • 2min
Exploring the Significance of Large Language Models and Cloud Computing in AI and ML
04:01 • 4min
Impact of Large Language Models on Data Science Jobs
08:05 • 5min
Development of a Course on Large Language Models
12:40 • 11min
Evolution of Language Models: From Neural Networks to Attention Mechanisms
23:18 • 6min
Evolution of Language Translation Mechanisms and Introduction of Transformers
29:20 • 5min
Technical Aspects of Transformers and LLMs
34:22 • 6min
Explaining Deep Learning with a Building Analogy
40:44 • 2min
Understanding Floor Numbering Systems and Introduction to Transformers and LLMs
42:31 • 11min
Technical Details of Attention Mechanism in Transformers and LLMs
53:47 • 33min
Deep Dive into Transformers and LLMs
01:26:54 • 17min
Exploring the Technical Aspects of Transformers and Large Language Models
01:44:05 • 15min
Technical Intro to Transformers and LLMs
01:58:42 • 2min
The Power of Psychology in Personal Development
02:00:18 • 4min
Technical Overview of Transformers and LLMs
02:04:17 • 2min