Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Introduction
00:00 • 3min
Can We Use State-Based Models in Language Modeling?
02:52 • 2min
Using State-Spaced Models in Language Modeling
05:19 • 3min
Using State Space Models in Machine Learning
07:57 • 2min
Using State Space Models to Model Long Sequences
10:23 • 3min
Can SSMs Do More Than Transformers?
13:19 • 3min
Can You Recall a Token?
16:37 • 2min
Is Hippo Campus a Good Idea?
18:14 • 2min
Hungry Hippos - What Are the High Level Ideas Behind This H3 Layer?
20:41 • 3min
Using Multiplier Interactions in RNN Modeling
23:14 • 2min
How Do You Measure the Extentivity of a Toy Task?
25:30 • 3min
H Three Language Models
28:25 • 3min
Is the Attention Layers Losing the Long Context Scaling Properties?
31:04 • 3min
Are Attention Layers Doing Great Things?
33:36 • 3min
What Are Some Product Applications?
36:13 • 3min
Is There a Future for AI Workflows?
39:00 • 3min