Limited Evaluation Evolutionary Optimization of Large Neural Networks
June 26, 2018 Β· Entered Twilight Β· π Deutsche Jahrestagung fΓΌr KΓΌnstliche Intelligenz
"Last commit was 7.0 years ago (β₯5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: .gitignore, README.md, baseline.py, baseline_cnn.py, evaluation.py, gpuea.py, gpuea_cnn.py, launch.py, launch.sh, lib.py
Authors
Jonas Prellberg, Oliver Kramer
arXiv ID
1806.09819
Category
cs.NE: Neural & Evolutionary
Citations
8
Venue
Deutsche Jahrestagung fΓΌr KΓΌnstliche Intelligenz
Repository
https://github.com/jprellberg/gpuea
Last Checked
1 month ago
Abstract
Stochastic gradient descent is the most prevalent algorithm to train neural networks. However, other approaches such as evolutionary algorithms are also applicable to this task. Evolutionary algorithms bring unique trade-offs that are worth exploring, but computational demands have so far restricted exploration to small networks with few parameters. We implement an evolutionary algorithm that executes entirely on the GPU, which allows to efficiently batch-evaluate a whole population of networks. Within this framework, we explore the limited evaluation evolutionary algorithm for neural network training and find that its batch evaluation idea comes with a large accuracy trade-off. In further experiments, we explore crossover operators and find that unprincipled random uniform crossover performs extremely well. Finally, we train a network with 92k parameters on MNIST using an EA and achieve 97.6 % test accuracy compared to 98 % test accuracy on the same network trained with Adam. Code is available at https://github.com/jprellberg/gpuea.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Neural & Evolutionary
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Progressive Growing of GANs for Improved Quality, Stability, and Variation
R.I.P.
π»
Ghosted
Learning both Weights and Connections for Efficient Neural Networks
R.I.P.
π»
Ghosted
LSTM: A Search Space Odyssey
R.I.P.
π»
Ghosted
A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks
R.I.P.
π»
Ghosted