Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability (Sparse Autoencoders)

Machine Learning Street Talk (MLST)

00:00

Exploring Sparse Autoencoders in Neural Networks

This chapter examines the role of sparse autoencoders in enhancing neural network interpretability and recognizing multilingual features. It highlights the challenges of mechanistic interpretability and the gap in understanding model inner workings while discussing ongoing research and advancements in the field.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app