On the Powerball Method for Optimization
March 24, 2016 ยท Declared Dead ยท ๐ IEEE Control Systems Letters
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Ye Yuan, Mu Li, Jun Liu, Claire J. Tomlin
arXiv ID
1603.07421
Category
eess.SY: Systems & Control (EE)
Cross-listed
cs.LG,
math.OC
Citations
22
Venue
IEEE Control Systems Letters
Last Checked
1 month ago
Abstract
We propose a new method to accelerate the convergence of optimization algorithms. This method simply adds a power coefficient $ฮณ\in[0,1)$ to the gradient during optimization. We call this the Powerball method and analyze the convergence rate for the Powerball method for strongly convex functions. While theoretically the Powerball method is guaranteed to have a linear convergence rate in the same order of the gradient method, we show that empirically it significantly outperforms the gradient descent and Newton's method, especially during the initial iterations. We demonstrate that the Powerball method provides a $10$-fold speedup of the convergence of both gradient descent and L-BFGS on multiple real datasets.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Systems & Control (EE)
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey
R.I.P.
๐ป
Ghosted
Wireless Network Design for Control Systems: A Survey
R.I.P.
๐ป
Ghosted
Learning-based Model Predictive Control for Safe Exploration
R.I.P.
๐ป
Ghosted
Safety-Critical Model Predictive Control with Discrete-Time Control Barrier Function
R.I.P.
๐ป
Ghosted
Novel Multidimensional Models of Opinion Dynamics in Social Networks
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted