Generally Intelligent cover image

Episode 24: Jack Parker-Holder, DeepMind, on open-endedness, evolving agents and environments, online adaptation, and offline learning

Generally Intelligent

00:00

Is the Generational Training a Good Idea?

The way it works in with generations is that you're not actually copying the weights you just went to the installation loss. It's like a KL between the action distributions of the previous agent and your agent on the new data. I see, I see. That actually makes a lot of sense. And there's some other theory for why the generational training process might work better. This is a bit wild, but it may not just be about overfitting,. But for the weight copying, what I always felt was that weight copying kind of erases information in the network that you're copying to.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app