Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
Introduction
00:00 • 3min
Is There a Career Pathway That You Want to Be On?
02:44 • 2min
Scale Ability of Alignment Solutions
04:50 • 2min
Is There a Time Line Tour to a G I Development?
06:49 • 3min
Are You Hedging Against Specific Risks?
09:27 • 2min
Reward Modelling in Reenforcement Learning
11:10 • 5min
Recursive Reward Modelling
15:46 • 5min
Is the Human Being at the Lower Level of the Recursion?
20:44 • 3min
Is There a Misalignment?
24:10 • 2min
Is There a Way to Measure Degree of Misalignment?
25:54 • 5min
The Balance Between Experimentation and Theory in the Alignment Community?
30:32 • 2min
I Don't Think Wire Heading Is That Widely Understandable Outside of an Allignment Research
32:41 • 4min
Is There a Scale Gap in Access to Large Models?
36:22 • 3min
Scaling Laws for Language Models
39:04 • 4min
Gpty 3
43:17 • 2min
What's Going on With G Pty Three?
45:38 • 2min
Reenforcementing From Human Feedback
47:14 • 2min
The Uncanny Valley of Capabilities and Alignment
49:08 • 2min
Is There a Competent Approach to Alinemend?
50:54 • 2min
Is Machine Learning a Soft Engineering Problem?
52:42 • 2min
A G I MVP
54:41 • 2min
How Do You Solve a Language Model That Produces a Harmful Output?
56:11 • 2min
How to Alline a System or Arbitrarity Powerful
58:27 • 3min
Is There Room for Totally Novel Approaches to This?
01:01:35 • 2min
Are You Hiring Reset Engineers?
01:03:21 • 2min