Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

CHAPTER

Unraveling Neural Networks

This chapter explores mechanistic interpretability, aiming to decode the complexities of neural networks viewed as black boxes. It highlights the challenges researchers face, such as superposition and the distinct ways models process information compared to humans. The discussion advocates for empirical approaches to understanding models' inner workings, emphasizing curiosity and the potential for breakthroughs in AI comprehension.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner