

Collin Burns On Discovering Latent Knowledge In Language Models Without Supervision
16 snips Jan 17, 2023
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63
Introduction
00:00 • 2min
The Five Seconds Pure Luck Is an Important Part of the World Record
01:36 • 3min
How Did You Become a Deep Learning Researcher?
04:36 • 2min
Is AI Really Important?
06:48 • 4min
Is There a Future for NLP?
10:56 • 2min
How Much Energy and Effort Will You Put Into Getting Superhuman Level in Math?
12:48 • 2min
Is Blenderbot a Good Benchmark for AGI?
14:50 • 2min
A Different AI to AI Research, What Do You Think?
17:14 • 2min
Automating AI Research
18:46 • 3min
Do You Think D.T.K. Will Have People Like Using AI to Produce New Ideas?
21:41 • 2min
How to Make Language Models Truthful
23:34 • 3min
Do Models Deliberately Lilt?
26:08 • 2min
Are You Good at Diplomacy?
28:25 • 2min
How to Train a Model to Do Open-End Deductions in Real-Time
29:57 • 2min
How to Discover Later Knowledge in the Lingered Models
32:22 • 2min
How Do We Train Normal Language Models?
34:17 • 3min
Is There a Difference Between Predicting and Not Prompting?
37:08 • 2min
Unsupervised Modeling
38:44 • 5min
Using Logical Consistency to Predict Next Tokens
43:19 • 2min
Logical Consistency Properties for Sentiment
45:12 • 2min
The Hidden State of the Last Token
47:12 • 2min
How to Extract a Hidden State in a 3D Model?
49:32 • 2min
Logistic Regression for Class?
51:41 • 2min
The Uh Local Optima That Gets High Accuracy on All Kinds of Questions
53:25 • 2min
Is There a Model of the World That Says the Truth?
55:38 • 2min
The Key Findings of Zero Shot Prompting
58:08 • 3min
The Conscious Consistent Church Method Is Better Than Zero Shot Prompting
01:00:39 • 4min
The Hidden State Is More Robust Than the Outputs
01:04:14 • 2min
The Main Level Takeaways From Your Paper
01:06:26 • 2min
Is Elon Musk Really Superhuman?
01:08:41 • 2min
Is There a Difference Between Superhuman Outputs and AI?
01:10:51 • 2min
Is a Model Like RL Breaking the Law or Not?
01:13:09 • 2min
Are You Breaking the Law?
01:15:13 • 3min
Is GPTN Dangerous?
01:17:53 • 2min
Rl From Human Feedback Is a Good Idea?
01:19:59 • 2min
Is AI Feedback a Better Way to Evaluate Models?
01:21:48 • 4min
AlphaGo and MuZero Are Good Alignment Isn't It?
01:25:20 • 2min
Open AI
01:27:11 • 3min
AI Alignment
01:30:12 • 3min
Do We Need Models to Maximize Profit Subjects?
01:33:04 • 2min
Alignment Problems
01:35:26 • 2min
How Similar Are Current Deep Learning Systems to GPTN?
01:37:00 • 5min
Using Unsupervised Models to Predict Future Text
01:41:53 • 2min
Is This Input True or False?
01:44:14 • 3min
Is It a Private Taper?
01:46:49 • 3min
Recovering the Truth in the Model
01:50:07 • 4min
Is There a Difference Between Constraints and the Truth?
01:53:44 • 5min
The Misaligned AI System
01:58:37 • 2min
Getting a Model to Say the Truth
02:00:53 • 2min
Is the Misaligned AI System Lieing?
02:02:24 • 2min
How to Train a Probation Better on the Misaligned AI
02:04:09 • 2min
I've Been Like Oh Yeah Ask Him if I Have a Sandwich and Not Ask Him
02:05:48 • 4min
Are You Egregiously Breaking the Law?
02:09:33 • 2min
Is Your Paper Doing What It's Not Doing Right?
02:11:15 • 2min
The Paper Doesn't Show the Complete Final Robust Method for This Problem
02:13:13 • 2min
Unsupervised Models - How to Find the Truth in a Linear Way
02:15:22 • 2min
How Do We Distinguish Between the Truth and the Misaligned System?
02:17:27 • 4min
Using a Prompt to Optimize the Model Class - Is This True or False?
02:21:01 • 2min
How Do Humans Do This or What Does That Even Mean?
02:23:25 • 2min
Thinking About Future AI Systems
02:25:11 • 2min
Having Access to the Biggest Models Is Important
02:26:46 • 2min
Is This the Most Important Problem That We're Currently Facing?
02:28:47 • 3min
Deep Learning Is Just Not That Deep
02:32:04 • 3min