The Thesis Review cover image

[23] Simon Du - Gradient Descent for Non-convex Problems in Modern Machine Learning

The Thesis Review

CHAPTER

Gradient Descent Theories in Neural Networks

This chapter explores the theoretical foundations of gradient descent and its various forms, emphasizing their relevance in training complex neural networks. It highlights the importance of understanding gradient flow, optimization gaps, and the role of ordinary differential equations in analyzing the dynamics of non-convex optimization.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner