Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions
May 02, 2016 ยท Declared Dead ยท ๐ Information Technology Convergence and Services
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Ioannis Panageas, Georgios Piliouras
arXiv ID
1605.00405
Category
math.DS
Cross-listed
cs.LG
Citations
154
Venue
Information Technology Convergence and Services
Last Checked
1 month ago
Abstract
Given a non-convex twice differentiable cost function f, we prove that the set of initial conditions so that gradient descent converges to saddle points where \nabla^2 f has at least one strictly negative eigenvalue has (Lebesgue) measure zero, even for cost functions f with non-isolated critical points, answering an open question in [Lee, Simchowitz, Jordan, Recht, COLT2016]. Moreover, this result extends to forward-invariant convex subspaces, allowing for weak (non-globally Lipschitz) smoothness assumptions. Finally, we produce an upper bound on the allowable step-size.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ math.DS
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Linearly-Recurrent Autoencoder Networks for Learning Dynamics
R.I.P.
๐ป
Ghosted
Eigendecompositions of Transfer Operators in Reproducing Kernel Hilbert Spaces
R.I.P.
๐ป
Ghosted
From rate distortion theory to metric mean dimension: variational principle
R.I.P.
๐ป
Ghosted
Double variational principle for mean dimension
R.I.P.
๐ป
Ghosted
Discovering conservation laws from data for control
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted