Changelog Master Feed cover image

Evaluating models without test data (Practical AI #194)

Changelog Master Feed

NOTE

Is There a Theory for Convex Optimization?

When you have a model and you're trying to train it, should you be trying to optimize the hyperparameters or should you be adding more data? You can think of it in a sort of very crude sense. And we know that a lot of these large language models, like BERT, for example, are just not properly converged. There are a large number of layers that are simply undertrained.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner