Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions

May 02, 2016 ยท Declared Dead ยท ๐Ÿ› Information Technology Convergence and Services

๐Ÿ‘ป CAUSE OF DEATH: Ghosted
No code link whatsoever

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Ioannis Panageas, Georgios Piliouras arXiv ID 1605.00405 Category math.DS Cross-listed cs.LG Citations 154 Venue Information Technology Convergence and Services Last Checked 1 month ago
Abstract
Given a non-convex twice differentiable cost function f, we prove that the set of initial conditions so that gradient descent converges to saddle points where \nabla^2 f has at least one strictly negative eigenvalue has (Lebesgue) measure zero, even for cost functions f with non-isolated critical points, answering an open question in [Lee, Simchowitz, Jordan, Recht, COLT2016]. Moreover, this result extends to forward-invariant convex subspaces, allowing for weak (non-globally Lipschitz) smoothness assumptions. Finally, we produce an upper bound on the allowable step-size.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” math.DS

Died the same way โ€” ๐Ÿ‘ป Ghosted