"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis cover image

Untangling Neural Network Mechanisms: Goodfire's Lee Sharkey on Parameter-based Interpretability

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

00:00

Understanding Neural Networks through Stacked Autoencoders

This chapter explores the functionality of Stacked Autoencoders (SAEs) in neural networks, highlighting their importance in parameter-based interpretability. The chapter explains how SAEs manage complex representations through a wide layer and focuses on the training process that balances sparsity with model behavior to identify key features.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app