AXRP - the AI X-risk Research Podcast cover image

35 - Peter Hase on LLM Beliefs and Easy-to-Hard Generalization

AXRP - the AI X-risk Research Podcast

00:00

Understanding Residual Layers and Information Flow in Language Models

This chapter examines the significance of residual layers in transformer architectures and how they contribute to information processing during a forward pass. It also discusses the implications of layer swapping and ongoing research aimed at enhancing model editing techniques.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app