Active Domain Randomization

April 09, 2019 ยท Entered Twilight ยท ๐Ÿ› Conference on Robot Learning

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, adr.gif, common, experiments, real_robot.py, real_robot_torquesweep.py, scripts, setup.py, slurm.sh, tests

Authors Bhairav Mehta, Manfred Diaz, Florian Golemo, Christopher J. Pal, Liam Paull arXiv ID 1904.04762 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.RO Citations 310 Venue Conference on Robot Learning Repository https://github.com/montrealrobotics/active-domainrand โญ 101 Last Checked 1 month ago
Abstract
Domain randomization is a popular technique for improving domain transfer, often used in a zero-shot setting when the target domain is unknown or cannot easily be used for training. In this work, we empirically examine the effects of domain randomization on agent generalization. Our experiments show that domain randomization may lead to suboptimal, high-variance policies, which we attribute to the uniform sampling of environment parameters. We propose Active Domain Randomization, a novel algorithm that learns a parameter sampling strategy. Our method looks for the most informative environment variations within the given randomization ranges by leveraging the discrepancies of policy rollouts in randomized and reference environment instances. We find that training more frequently on these instances leads to better overall agent generalization. Our experiments across various physics-based simulated and real-robot tasks show that this enhancement leads to more robust, consistent policies.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning