

#69 DR. THOMAS LUX - Interpolation of Sparse High-Dimensional Data
18 snips Mar 12, 2022
Dr. Thomas Lux, a research scientist at Meta in Silicon Valley, dives deep into the geometry behind machine learning. He discusses the unique advantages of neural networks over classical methods for high-dimensional data interpolation. Lux explains how neural networks excel at tasks like image recognition by effectively reducing dimensions and ignoring irrelevant data. He explores the challenges of placing basis functions and the importance of data density. Their ability to focus on crucial input regions reveals why they outperform traditional algorithms.
AI Snips
Chapters
Transcript
Episode notes
Geometric Lens of Machine Learning
- Supervised machine learning can be viewed geometrically, with training data as points in Euclidean space.
- The goal is to predict a function's value at these points, often using neural networks.
Neural Networks' Advantage
- Neural networks excel at image recognition due to their ability to perform nonlinear dimension reduction.
- They concentrate approximation power on important function parts, ignoring less relevant input areas.
Delaunay Triangulation Explained
- Delaunay triangulation generalizes linear interpolation to higher dimensions by connecting data points to form simplices.
- These simplices create a piecewise linear interpolant, useful for approximating functions in higher dimensions.