
[23] Simon Du - Gradient Descent for Non-convex Problems in Modern Machine Learning
The Thesis Review
Gradient Descent Theories in Neural Networks
This chapter explores the theoretical foundations of gradient descent and its various forms, emphasizing their relevance in training complex neural networks. It highlights the importance of understanding gradient flow, optimization gaps, and the role of ordinary differential equations in analyzing the dynamics of non-convex optimization.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.