Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability (Sparse Autoencoders)

Machine Learning Street Talk (MLST)

00:00

Exploring Sparse Autoencoders in Neural Networks

This chapter examines the role of sparse autoencoders in enhancing neural network interpretability and recognizing multilingual features. It highlights the challenges of mechanistic interpretability and the gap in understanding model inner workings while discussing ongoing research and advancements in the field.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app