Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

00:00

Unraveling Neural Networks

This chapter explores mechanistic interpretability, aiming to decode the complexities of neural networks viewed as black boxes. It highlights the challenges researchers face, such as superposition and the distinct ways models process information compared to humans. The discussion advocates for empirical approaches to understanding models' inner workings, emphasizing curiosity and the potential for breakthroughs in AI comprehension.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app