AXRP - the AI X-risk Research Podcast cover image

21 - Interpretability for Engineers with Stephen Casper

AXRP - the AI X-risk Research Podcast

00:00

The Future of Deep Learning

I think the deep learning field has shown to be one that's guided by empirical progress, much more than theoretical progress. One could argue that interpretability for AI safety has been quite popular since maybe 2017 and people are saying very similar things now. I have no argument in theory for why we should never expect this. But there are some reasons to be a little bit doubtful that we might be able to kind of, you know, basic science our way into understanding things in a very, very useful and very, very rigorous way.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app