Affordance Learning for End-to-End Visuomotor Robot Control

March 10, 2019 Β· Declared Dead Β· πŸ› IEEE/RJS International Conference on Intelligent RObots and Systems

πŸ‘» CAUSE OF DEATH: Ghosted
No code link whatsoever

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Aleksi HΓ€mΓ€lΓ€inen, Karol Arndt, Ali Ghadirzadeh, Ville Kyrki arXiv ID 1903.04053 Category cs.RO: Robotics Cross-listed cs.LG Citations 50 Venue IEEE/RJS International Conference on Intelligent RObots and Systems Last Checked 1 month ago
Abstract
Training end-to-end deep robot policies requires a lot of domain-, task-, and hardware-specific data, which is often costly to provide. In this work, we propose to tackle this issue by employing a deep neural network with a modular architecture, consisting of separate perception, policy, and trajectory parts. Each part of the system is trained fully on synthetic data or in simulation. The data is exchanged between parts of the system as low-dimensional latent representations of affordances and trajectories. The performance is then evaluated in a zero-shot transfer scenario using Franka Panda robot arm. Results demonstrate that a low-dimensional representation of scene affordances extracted from an RGB image is sufficient to successfully train manipulator policies. We also introduce a method for affordance dataset generation, which is easily generalizable to new tasks, objects and environments, and requires no manual pixel labeling.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Robotics

Died the same way β€” πŸ‘» Ghosted