Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability (Sparse Autoencoders)

Machine Learning Street Talk (MLST)

CHAPTER

Exploring Sparse Autoencoders in Neural Networks

This chapter examines the role of sparse autoencoders in enhancing neural network interpretability and recognizing multilingual features. It highlights the challenges of mechanistic interpretability and the gap in understanding model inner workings while discussing ongoing research and advancements in the field.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner