
Evaluating models without test data (Practical AI #194)
Changelog Master Feed
Is There a Theory for Convex Optimization?
When you have a model and you're trying to train it, should you be trying to optimize the hyperparameters or should you be adding more data? You can think of it in a sort of very crude sense. And we know that a lot of these large language models, like BERT, for example, are just not properly converged. There are a large number of layers that are simply undertrained.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.