6min chapter

Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

CHAPTER

How Neural Networks Learn Group Operations

weight decay is like an inductive bias essentially to tell the model to reduce its complexity which is a pressure to generalize. If it wasn't for that then that wouldn't happen so in the experiments I ran if you don't have weight decay it will just keep memorizing infinitely far. Each run of the model learns some combination of these circuits for the different representations each time. The exact number varies and which ones it learns is seemingly random each time which suggests that all time models lie to you obviously but if we're trying to reason about real networks looking at this work might suggest the explanation.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode