The Inside View cover image

Jesse Hoogland on Developmental Interpretability and Singular Learning Theory

The Inside View

00:00

The Effects of Face Transitions on Neural Networks

In a magnet, you only need an infinitesimally small perturbation to that system in order to point it in a certain direction. And so if you bring this back to neural networks, if the weights are slightly different, the overall model could be like the sepsilon or not. So there are these toy models of superposition studied by anthropic. You can see that the way compresses data involves this kind of symmetry breaking and selects one solution out of phase transition.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app