

126 - Optimizing Continuous Prompts for Generation, with Lisa Li
5 snips May 24, 2021
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Introduction
00:00 • 2min
How Prefix Tuning Compares to Lightweight Fine-Tuning
02:12 • 2min
How Prefix Tuning Compares With Adapter Tuning
04:28 • 5min
The Difference Between Manual and GPT-3 Prompts
09:23 • 5min
How to Optimize Prefix Parameters for Training
14:05 • 3min
Why Optimization Is Stable?
16:58 • 2min
The Differences Between Overparameterization and Adapter Tuning
18:42 • 3min
Table to Text and Summarization Tasks
21:56 • 2min
The Differences Between Table to Text and Summarization
24:03 • 2min
Prefix Tuning Outperforms Fine Tuning
25:55 • 4min
The Differences Between Prefix Tuning and Adapter Tuning
29:43 • 2min
The General Trends in the Results of Summarization Experiments
32:00 • 2min
Prefix Tuning on Encoder With Decoder
33:52 • 3min
Prefix Tuning: An Alternative to Fine Tuning
37:21 • 3min
The Importance of Prefix Tuning in Synchronization Tasks
40:41 • 5min
How to Transfer Prefixes Between Models
45:40 • 2min