R.I.P.
๐ป
Ghosted
Efficient Quality-Diversity Optimization through Diverse Quality Species
April 14, 2023 ยท Entered Twilight ยท ๐ GECCO Companion
Repo contents: .gitignore, CVT, README.md, main.py, neat_rl, requirements.txt, setup.py
Authors
Ryan Wickman, Bibek Poudel, Michael Villarreal, Xiaofei Zhang, Weizi Li
arXiv ID
2304.07425
Category
cs.LG: Machine Learning
Cross-listed
cs.AI,
cs.NE
Citations
7
Venue
GECCO Companion
Repository
https://github.com/rwickman/NEAT_RL
โญ 4
Last Checked
1 month ago
Abstract
A prevalent limitation of optimizing over a single objective is that it can be misguided, becoming trapped in local optimum. This can be rectified by Quality-Diversity (QD) algorithms, where a population of high-quality and diverse solutions to a problem is preferred. Most conventional QD approaches, for example, MAP-Elites, explicitly manage a behavioral archive where solutions are broken down into predefined niches. In this work, we show that a diverse population of solutions can be found without the limitation of needing an archive or defining the range of behaviors in advance. Instead, we break down solutions into independently evolving species and use unsupervised skill discovery to learn diverse, high-performing solutions. We show that this can be done through gradient-based mutations that take on an information theoretic perspective of jointly maximizing mutual information and performance. We propose Diverse Quality Species (DQS) as an alternative to archive-based QD algorithms. We evaluate it over several simulated robotic environments and show that it can learn a diverse set of solutions from varying species. Furthermore, our results show that DQS is more sample-efficient and performant when compared to other QD algorithms. Relevant code and hyper-parameters are available at: https://github.com/rwickman/NEAT_RL.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted