"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis cover image

Untangling Neural Network Mechanisms: Goodfire's Lee Sharkey on Parameter-based Interpretability

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

00:00

Decomposing Neural Networks for Interpretability

This chapter explores the process of breaking down neural network computations into simpler, atomic units to enhance analysis and interpretation. It emphasizes the importance of computational simplicity and loss functions that support minimality in training neural networks. Through discussions of theoretical frameworks and practical examples, the chapter addresses the complexities of model parameters and highlights the significance of maintaining fidelity while navigating high-dimensional learning challenges.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app