Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

00:00

Exploring Supermasks and Neural Activations

This chapter examines supermasks and their role in managing sparse subnetworks within models like transformers and CNNs, particularly in the context of catastrophic forgetting and continual learning. It also investigates the function of specific neurons in language models, focusing on their ability to disambiguate linguistic contexts and the challenges of proving monosemantic neuron specificity in language detection.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app