Semantic Part Detection via Matching: Learning to Generalize to Novel Viewpoints from Limited Training Data

November 28, 2018 ยท Entered Twilight ยท ๐Ÿ› IEEE International Conference on Computer Vision

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 6.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .DS_Store, FeatureExtractor_torch.py, MaximalCliqueAlgorithm, Pascal3dPlus.py, README.md, aggregate_3d_wPars.py, assign_synthetic_img.py, calculate_mAP.py, crop_synthetic_img.py, eval, extractLayerFeat.py, featPickle2featCell.py, get_vp_examples.py, refine_pool3.py, solve_nearest.py, solve_nearest_score.py, util.py, vertex_id_picker.py

Authors Yutong Bai, Qing Liu, Lingxi Xie, Weichao Qiu, Yan Zheng, Alan Yuille arXiv ID 1811.11823 Category cs.CV: Computer Vision Citations 10 Venue IEEE International Conference on Computer Vision Repository https://github.com/ytongbai/SemanticPartDetection โญ 10 Last Checked 1 month ago
Abstract
Detecting semantic parts of an object is a challenging task in computer vision, particularly because it is hard to construct large annotated datasets due to the difficulty of annotating semantic parts. In this paper we present an approach which learns from a small training dataset of annotated semantic parts, where the object is seen from a limited range of viewpoints, but generalizes to detect semantic parts from a much larger range of viewpoints. Our approach is based on a matching algorithm for finding accurate spatial correspondence between two images, which enables semantic parts annotated on one image to be transplanted to another. In particular, this enables images in the training dataset to be matched to a virtual 3D model of the object (for simplicity, we assume that the object viewpoint can be estimated by standard techniques). Then a clustering algorithm is used to annotate the semantic parts of the 3D virtual model. This virtual 3D model can be used to synthesize annotated images from a large range of viewpoint. These can be matched to images in the test set, using the same matching algorithm, to detect semantic parts in novel viewpoints of the object. Our algorithm is very simple, intuitive, and contains very few parameters. We evaluate our approach in the car subclass of the VehicleSemanticPart dataset. We show it outperforms standard deep network approaches and, in particular, performs much better on novel viewpoints. For facilitating the future research, code is available: https://github.com/ytongbai/SemanticPartDetection
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision