The Thesis Review cover image

[23] Simon Du - Gradient Descent for Non-convex Problems in Modern Machine Learning

The Thesis Review

00:00

Gradient Descent Theories in Neural Networks

This chapter explores the theoretical foundations of gradient descent and its various forms, emphasizing their relevance in training complex neural networks. It highlights the importance of understanding gradient flow, optimization gaps, and the role of ordinary differential equations in analyzing the dynamics of non-convex optimization.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app