6min chapter

Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

CHAPTER

The Causal Link Between in-Context Learning and Induction Heads

induction heads are to my eyes the canonical example of an inference time algorithm stored in the model's weights that get supplied. I'm sure there's a bunch more that no one has yet found and yeah a lot of my model is the prompt engineering is just telling the model which of its circuits to activate. A particularly fun qualitative study was looking at soft induction heads heads that seem to be doing something induction e in other domains like a few-shot learning head on this random synthetic pattern recognition task. i really want a better and more scientific understanding of emergent why does that happen really understanding particularly notable case studies of it testing the hypothesis that it is driven by specific kinds of circuits like induction heads or

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode