The Inside View

Collin Burns On Discovering Latent Knowledge In Language Models Without Supervision

16 snips
Jan 17, 2023
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 2min
2
The Five Seconds Pure Luck Is an Important Part of the World Record
01:36 • 3min
3
How Did You Become a Deep Learning Researcher?
04:36 • 2min
4
Is AI Really Important?
06:48 • 4min
5
Is There a Future for NLP?
10:56 • 2min
6
How Much Energy and Effort Will You Put Into Getting Superhuman Level in Math?
12:48 • 2min
7
Is Blenderbot a Good Benchmark for AGI?
14:50 • 2min
8
A Different AI to AI Research, What Do You Think?
17:14 • 2min
9
Automating AI Research
18:46 • 3min
10
Do You Think D.T.K. Will Have People Like Using AI to Produce New Ideas?
21:41 • 2min
11
How to Make Language Models Truthful
23:34 • 3min
12
Do Models Deliberately Lilt?
26:08 • 2min
13
Are You Good at Diplomacy?
28:25 • 2min
14
How to Train a Model to Do Open-End Deductions in Real-Time
29:57 • 2min
15
How to Discover Later Knowledge in the Lingered Models
32:22 • 2min
16
How Do We Train Normal Language Models?
34:17 • 3min
17
Is There a Difference Between Predicting and Not Prompting?
37:08 • 2min
18
Unsupervised Modeling
38:44 • 5min
19
Using Logical Consistency to Predict Next Tokens
43:19 • 2min
20
Logical Consistency Properties for Sentiment
45:12 • 2min
21
The Hidden State of the Last Token
47:12 • 2min
22
How to Extract a Hidden State in a 3D Model?
49:32 • 2min
23
Logistic Regression for Class?
51:41 • 2min
24
The Uh Local Optima That Gets High Accuracy on All Kinds of Questions
53:25 • 2min
25
Is There a Model of the World That Says the Truth?
55:38 • 2min
26
The Key Findings of Zero Shot Prompting
58:08 • 3min
27
The Conscious Consistent Church Method Is Better Than Zero Shot Prompting
01:00:39 • 4min
28
The Hidden State Is More Robust Than the Outputs
01:04:14 • 2min
29
The Main Level Takeaways From Your Paper
01:06:26 • 2min
30
Is Elon Musk Really Superhuman?
01:08:41 • 2min
31
Is There a Difference Between Superhuman Outputs and AI?
01:10:51 • 2min
32
Is a Model Like RL Breaking the Law or Not?
01:13:09 • 2min
33
Are You Breaking the Law?
01:15:13 • 3min
34
Is GPTN Dangerous?
01:17:53 • 2min
35
Rl From Human Feedback Is a Good Idea?
01:19:59 • 2min
36
Is AI Feedback a Better Way to Evaluate Models?
01:21:48 • 4min
37
AlphaGo and MuZero Are Good Alignment Isn't It?
01:25:20 • 2min
38
Open AI
01:27:11 • 3min
39
AI Alignment
01:30:12 • 3min
40
Do We Need Models to Maximize Profit Subjects?
01:33:04 • 2min
41
Alignment Problems
01:35:26 • 2min
42
How Similar Are Current Deep Learning Systems to GPTN?
01:37:00 • 5min
43
Using Unsupervised Models to Predict Future Text
01:41:53 • 2min
44
Is This Input True or False?
01:44:14 • 3min
45
Is It a Private Taper?
01:46:49 • 3min
46
Recovering the Truth in the Model
01:50:07 • 4min
47
Is There a Difference Between Constraints and the Truth?
01:53:44 • 5min
48
The Misaligned AI System
01:58:37 • 2min
49
Getting a Model to Say the Truth
02:00:53 • 2min
50
Is the Misaligned AI System Lieing?
02:02:24 • 2min
51
How to Train a Probation Better on the Misaligned AI
02:04:09 • 2min
52
I've Been Like Oh Yeah Ask Him if I Have a Sandwich and Not Ask Him
02:05:48 • 4min
53
Are You Egregiously Breaking the Law?
02:09:33 • 2min
54
Is Your Paper Doing What It's Not Doing Right?
02:11:15 • 2min
55
The Paper Doesn't Show the Complete Final Robust Method for This Problem
02:13:13 • 2min
56
Unsupervised Models - How to Find the Truth in a Linear Way
02:15:22 • 2min
57
How Do We Distinguish Between the Truth and the Misaligned System?
02:17:27 • 4min
58
Using a Prompt to Optimize the Model Class - Is This True or False?
02:21:01 • 2min
59
How Do Humans Do This or What Does That Even Mean?
02:23:25 • 2min
60
Thinking About Future AI Systems
02:25:11 • 2min
61
Having Access to the Biggest Models Is Important
02:26:46 • 2min
62
Is This the Most Important Problem That We're Currently Facing?
02:28:47 • 3min
63
Deep Learning Is Just Not That Deep
02:32:04 • 3min