Doduo: Learning Dense Visual Correspondence from Unsupervised Semantic-Aware Flow

September 26, 2023 ยท Entered Twilight ยท ๐Ÿ› IEEE International Conference on Robotics and Automation

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .env.example, .gitignore, .pre-commit-config.yaml, LICENSE, README.md, configs, data, env.yaml, notebooks, pyproject.toml, requirements.txt, setup.py, src

Authors Zhenyu Jiang, Hanwen Jiang, Yuke Zhu arXiv ID 2309.15110 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.RO Citations 8 Venue IEEE International Conference on Robotics and Automation Repository https://github.com/ut-austin-rpl/Doduo โญ 44 Last Checked 8 days ago
Abstract
Dense visual correspondence plays a vital role in robotic perception. This work focuses on establishing the dense correspondence between a pair of images that captures dynamic scenes undergoing substantial transformations. We introduce Doduo to learn general dense visual correspondence from in-the-wild images and videos without ground truth supervision. Given a pair of images, it estimates the dense flow field encoding the displacement of each pixel in one image to its corresponding pixel in the other image. Doduo uses flow-based warping to acquire supervisory signals for the training. Incorporating semantic priors with self-supervised flow training, Doduo produces accurate dense correspondence robust to the dynamic changes of the scenes. Trained on an in-the-wild video dataset, Doduo illustrates superior performance on point-level correspondence estimation over existing self-supervised correspondence learning baselines. We also apply Doduo to articulation estimation and zero-shot goal-conditioned manipulation, underlining its practical applications in robotics. Code and additional visualizations are available at https://ut-austin-rpl.github.io/Doduo
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision