Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Introduction
00:00 • 3min
Canavera, Col., Is That Right?
02:32 • 3min
Is There a Connection Between the Former Languages and the Shomsky Hiry Kind of Stuff?
05:05 • 2min
How to Characterize Inducto Bias
07:01 • 2min
A, Ye Oge, So Let's Look at the Haraco Structure of Natural Language
08:52 • 2min
How to Do This in Outcast?
10:30 • 2min
Is There a Bounded Depth on the Stack?
12:03 • 2min
The Limitation of Unbounded Depth Transformers
13:56 • 2min
How Much Memory Do You Need to Process This Dyke Language?
15:43 • 2min
Is There a Difference in the Precision?
17:14 • 3min
The Differences Between Recurrent Mechanism and Self Attention Mechanism
20:20 • 3min
Aranan, I Agree With Everything.
22:53 • 4min
The Intuition of the Self Attention Network
26:33 • 3min
The Scaler Position Incodi, Is Important for Former Languages?
29:30 • 3min
Generalization to Longer Sequence Lents?
32:12 • 1min
Why Isn't the Absolute Position Incoding So Important?
33:41 • 2min