Online Learning of a Memory for Learning Rates

September 20, 2017 ยท Entered Twilight ยท ๐Ÿ› IEEE International Conference on Robotics and Automation

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, LICENSE, README.md, meta_learning, notebooks, scripts, setup.py, third_party, ubash

Authors Franziska Meier, Daniel Kappler, Stefan Schaal arXiv ID 1709.06709 Category cs.LG: Machine Learning Citations 21 Venue IEEE International Conference on Robotics and Automation Repository https://github.com/fmeier/online-meta-learning โญ 26 Last Checked 1 month ago
Abstract
The promise of learning to learn for robotics rests on the hope that by extracting some information about the learning process itself we can speed up subsequent similar learning tasks. Here, we introduce a computationally efficient online meta-learning algorithm that builds and optimizes a memory model of the optimal learning rate landscape from previously observed gradient behaviors. While performing task specific optimization, this memory of learning rates predicts how to scale currently observed gradients. After applying the gradient scaling our meta-learner updates its internal memory based on the observed effect its prediction had. Our meta-learner can be combined with any gradient-based optimizer, learns on the fly and can be transferred to new optimization tasks. In our evaluations we show that our meta-learning algorithm speeds up learning of MNIST classification and a variety of learning control tasks, either in batch or online learning settings.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning