Machine Learning Street Talk (MLST) cover image

Neel Nanda - Mechanistic Interpretability

Machine Learning Street Talk (MLST)

CHAPTER

Exploring Complexity in Language Models

This chapter discusses the implications of larger model and data sizes in language models and their ability to perform complex tasks, such as coding and solving logic puzzles. The conversation examines whether these models memorize data or develop a deeper understanding of language and world dynamics, highlighting insights from the Othello paper and exploring the intricacies of neural network behavior and interpretability. It critiques existing approaches and theorizes about the nature of AI reasoning, emphasizing the challenges of model learning and representation.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner