

Episode 26: Developing and Training LLMs From Scratch
May 15, 2024
Sebastian Raschka discusses developing and training large language models (LLMs) from scratch, covering topics like prompt engineering, fine-tuning, and RAG systems. They explore the skills, resources, and hardware needed, the lifecycle of LLMs, live coding to create a spam classifier, and the importance of hands-on experience. They also touch on using PyTorch Lightning and fabric for managing large models, and reveal insights on techniques in natural language processing models and evaluating LLMs for classification problems.
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9
Introduction
00:00 • 2min
Exploring backgrounds and interests in AI and machine learning
02:01 • 4min
Exploring the Power of Large Language Models
06:15 • 30min
Exploring Skills and Challenges in Working with LLMs
36:23 • 4min
Implementing Deep Learning Models from Scratch
40:51 • 21min
Exploring Techniques in Natural Language Processing Models
01:01:48 • 4min
Evolution of Language Model Techniques
01:05:32 • 30min
Evaluating Large Language Models for Classification Problems
01:35:35 • 5min
Navigating Commitments and Prioritizing Excitement in Generative AI
01:40:14 • 11min