

ChatGPT and InstructGPT: Aligning Language Models to Human Intention
14 snips Jan 18, 2023
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
Introduction
00:00 • 2min
ML Observability and Alignment - Part 1 of 3
01:35 • 2min
The Main Problems With Large Language Models
03:42 • 2min
Is the Term Alignment Created by OpenAI?
05:50 • 2min
Instruct GPT Paper - What Is It?
07:22 • 2min
How Do You Train a Reward Model?
09:50 • 2min
How to Train a Reward Model for a Good Job
11:36 • 3min
Do You See Other Major Applications Just Skipping the First Step?
14:10 • 2min
How Did You Come Up With the Idea of Instructing GPT?
16:07 • 2min
The Nature of Prompt Engineering Will Change Over Time
18:35 • 2min
Are There Other Major Benefits of LHF?
20:35 • 4min
OpenAI
24:55 • 2min
Is There a Jury on Large Language Models?
26:43 • 2min
The Next Generation of Language Models Are Going to Be Really Really Powerful
28:30 • 2min
Is There a Future for Machine Learning?
30:45 • 3min
How Much Is the Reward Model the Most Important?
33:45 • 2min
Is LHF the Best Way to Fine Tune Language Models?
35:33 • 2min
Is There a Way to Evaluate a Powerful Model?
37:10 • 2min
The Long Term Alignment Research
38:43 • 2min
Is There Anything You Can Point to That's Not Really Great?
40:26 • 2min
Is That Part of the Actual Training or Is That Like the Fine Tuning?
42:36 • 3min
I'd Like to Give Chat and Instruct GPT a Try
45:50 • 2min