

Daniela and Dario Amodei on Anthropic
02:01:27
Anthropic's Focused Research Bet
- Anthropic was founded to make a focused research bet with a small, aligned team.
- They aim to build reliable, interpretable, and steerable AI systems.
Anthropic's Research Strategy
- Anthropic trains large-scale generative models and conducts safety research on them.
- Their research includes interpretability, alignment, societal impact, and scaling laws.
AI Existential Safety and Current Model Problems
- AI existential safety is a long-term motivation, but Anthropic focuses on current model problems.
- These problems, like open-endedness and unpredictability, offer insights into future AI safety.
Get the Snipd Podcast app to discover more snips from this episode
Get the app 1 chevron_right 2 chevron_right 3 chevron_right 4 chevron_right 5 chevron_right 6 chevron_right 7 chevron_right 8 chevron_right 9 chevron_right 10 chevron_right 11 chevron_right 12 chevron_right 13 chevron_right 14 chevron_right 15 chevron_right 16 chevron_right 17 chevron_right 18 chevron_right 19 chevron_right 20 chevron_right 21 chevron_right 22 chevron_right 23 chevron_right 24 chevron_right 25 chevron_right 26 chevron_right 27 chevron_right 28 chevron_right 29 chevron_right 30 chevron_right 31 chevron_right 32 chevron_right 33 chevron_right 34 chevron_right 35 chevron_right 36 chevron_right
Introduction
00:00 • 3min
How Did You Find Anthropic?
02:43 • 4min
Scaling to a G I
06:24 • 2min
Is There a Shared Vision?
07:55 • 3min
Is Existential Safety a Threat to Humanity?
11:10 • 3min
The Challenge of General, Open Endowed Models
14:14 • 5min
Is It Possible for a Language Model to Lies?
19:29 • 3min
Predictability and Surprisingness in Generative Models
22:45 • 5min
How Do You Make Safe Systems?
27:36 • 5min
Can Academia Contribute to Safety?
32:49 • 3min
The Boiler Plate Mission of Anthropics
36:17 • 2min
Scaling to Higher Levels of Intelligence
38:40 • 5min
Ai Systems of the Future Look Different From the Way They Look to Day
43:25 • 3min
Is Your Research in Papers Going to Change the World?
46:23 • 3min
Trying to Interpret Current Models?
49:53 • 6min
Interpretability Release - Tools and Vidios
55:36 • 4min
Is There a Space for Interpretability Research?
59:07 • 2min
How Do You Get Along With Your Brother and Sisterya?
01:01:06 • 4min
I'll Take a Stab at This Ad and Maybe, I'll Just Kind of Gesture in a Few Diffrent Directions Here
01:05:33 • 3min
The Impact of Artificial Intelligence
01:08:12 • 4min
Scaling Machine Learning Systems
01:11:49 • 5min
Scaling Is the Core Ingredient
01:16:52 • 1min
Will We Get All the Way to General Intelligence?
01:18:01 • 3min
Is There Less of Them?
01:20:56 • 2min
Anthropic Research
01:22:27 • 3min
What Is a Public Benefit Corporation?
01:25:36 • 2min
Anthropic Corporation
01:27:49 • 3min
Is There a Mass of Economic Windfalls From a I?
01:30:25 • 4min
Is There a Relationship Between Current Day Problems and Accidential Risk?
01:34:08 • 4min
Is There a Plan to Develop a G I Safely?
01:38:04 • 3min
The Role and Importance of Governments in Civil Society
01:41:29 • 4min
What Rolls Do You Expect to Be Hiring?
01:45:13 • 3min
Anthropic: Is It a Great Team?
01:48:31 • 4min
Why Do You Choose to Hire People From Outside the Field?
01:52:49 • 2min
The Future Impacts of Advanced Ai Systems
01:54:34 • 5min
The Best Place to Follow Philanthropic on Twitter
01:59:44 • 2min
Daniela and Dario Amodei join us to discuss Anthropic: a new AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.
Topics discussed in this episode include:
-Anthropic's mission and research strategy
-Recent research and papers by Anthropic
-Anthropic's structure as a "public benefit corporation"
-Career opportunities
You can find the page for the podcast here: https://futureoflife.org/2022/03/04/daniela-and-dario-amodei-on-anthropic/
Watch the video version of this episode here: https://www.youtube.com/watch?v=uAA6PZkek4A
Careers at Anthropic: https://www.anthropic.com/#careers
Anthropic's Transformer Circuits research: https://transformer-circuits.pub/
Follow Anthropic on Twitter: https://twitter.com/AnthropicAI
microCOVID Project: https://www.microcovid.org/
Follow Lucas on Twitter: https://twitter.com/lucasfmperry
Have any feedback about the podcast? You can share your thoughts here:
www.surveymonkey.com/r/DRBFZCT
Timestamps:
0:00 Intro
2:44 What was the intention behind forming Anthropic?
6:28 Do the founders of Anthropic share a similar view on AI?
7:55 What is Anthropic's focused research bet?
11:10 Does AI existential safety fit into Anthropic's work and thinking?
14:14 Examples of AI models today that have properties relevant to future AI existential safety
16:12 Why work on large scale models?
20:02 What does it mean for a model to lie?
22:44 Safety concerns around the open-endedness of large models
29:01 How does safety work fit into race dynamics to more and more powerful AI?
36:16 Anthropic's mission and how it fits into AI alignment
38:40 Why explore large models for AI safety and scaling to more intelligent systems?
43:24 Is Anthropic's research strategy a form of prosaic alignment?
46:22 Anthropic's recent research and papers
49:52 How difficult is it to interpret current AI models?
52:40 Anthropic's research on alignment and societal impact
55:35 Why did you decide to release tools and videos alongside your interpretability research
1:01:04 What is it like working with your sibling?
1:05:33 Inspiration around creating Anthropic
1:12:40 Is there an upward bound on capability gains from scaling current models?
1:18:00 Why is it unlikely that continuously increasing the number of parameters on models will lead to AGI?
1:21:10 Bootstrapping models
1:22:26 How does Anthropic see itself as positioned in the AI safety space?
1:25:35 What does being a public benefit corporation mean for Anthropic?
1:30:55 Anthropic's perspective on windfall profits from powerful AI systems
1:34:07 Issues with current AI systems and their relationship with long-term safety concerns
1:39:30 Anthropic's plan to communicate it's work to technical researchers and policy makers
1:41:28 AI evaluations and monitoring
1:42:50 AI governance
1:45:12 Careers at Anthropic
1:48:30 What it's like working at Anthropic
1:52:48 Why hire people of a wide variety of technical backgrounds?
1:54:33 What's a future you're excited about or hopeful for?
1:59:42 Where to find and follow Anthropic
This podcast is possible because of the support of listeners like you. If you found this conversation to be meaningful or valuable, consider supporting it directly by donating at futureoflife.org/donate. Contributions like yours make these conversations possible.