Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

CHAPTER

Exploring Supermasks and Neural Activations

This chapter examines supermasks and their role in managing sparse subnetworks within models like transformers and CNNs, particularly in the context of catastrophic forgetting and continual learning. It also investigates the function of specific neurons in language models, focusing on their ability to disambiguate linguistic contexts and the challenges of proving monosemantic neuron specificity in language detection.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner