Meta-Controller: Few-Shot Imitation of Unseen Embodiments and Tasks in Continuous Control

December 10, 2024 ยท Declared Dead ยท ๐Ÿ› Neural Information Processing Systems

๐Ÿ’€ CAUSE OF DEATH: 404 Not Found
Code link is broken/dead
Authors Seongwoong Cho, Donggyun Kim, Jinwoo Lee, Seunghoon Hong arXiv ID 2412.12147 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.RO Citations 2 Venue Neural Information Processing Systems Repository https://github.com/SeongwoongCho/meta-controller}{https://github.com/SeongwoongCho/meta-controller} Last Checked 1 month ago
Abstract
Generalizing across robot embodiments and tasks is crucial for adaptive robotic systems. Modular policy learning approaches adapt to new embodiments but are limited to specific tasks, while few-shot imitation learning (IL) approaches often focus on a single embodiment. In this paper, we introduce a few-shot behavior cloning framework to simultaneously generalize to unseen embodiments and tasks using a few (\emph{e.g.,} five) reward-free demonstrations. Our framework leverages a joint-level input-output representation to unify the state and action spaces of heterogeneous embodiments and employs a novel structure-motion state encoder that is parameterized to capture both shared knowledge across all embodiments and embodiment-specific knowledge. A matching-based policy network then predicts actions from a few demonstrations, producing an adaptive policy that is robust to over-fitting. Evaluated in the DeepMind Control suite, our framework termed \modelname{} demonstrates superior few-shot generalization to unseen embodiments and tasks over modular policy learning and few-shot IL approaches. Codes are available at \href{https://github.com/SeongwoongCho/meta-controller}{https://github.com/SeongwoongCho/meta-controller}.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning

Died the same way โ€” ๐Ÿ’€ 404 Not Found