Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation

October 16, 2018 ยท Entered Twilight ยท ๐Ÿ› Conference on Robot Learning

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, configs, scripts, src

Authors Gregory Kahn, Adam Villaflor, Pieter Abbeel, Sergey Levine arXiv ID 1810.07167 Category cs.RO: Robotics Cross-listed cs.AI, cs.LG Citations 21 Venue Conference on Robot Learning Repository https://github.com/gkahn13/CAPs โญ 33 Last Checked 1 month ago
Abstract
A general-purpose intelligent robot must be able to learn autonomously and be able to accomplish multiple tasks in order to be deployed in the real world. However, standard reinforcement learning approaches learn separate task-specific policies and assume the reward function for each task is known a priori. We propose a framework that learns event cues from off-policy data, and can flexibly combine these event cues at test time to accomplish different tasks. These event cue labels are not assumed to be known a priori, but are instead labeled using learned models, such as computer vision detectors, and then `backed up' in time using an action-conditioned predictive model. We show that a simulated robotic car and a real-world RC car can gather data and train fully autonomously without any human-provided labels beyond those needed to train the detectors, and then at test-time be able to accomplish a variety of different tasks. Videos of the experiments and code can be found at https://github.com/gkahn13/CAPs
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Robotics