Benchmarking and Error Diagnosis in Multi-Instance Pose Estimation

July 17, 2017 ยท Entered Twilight ยท ๐Ÿ› IEEE International Conference on Computer Vision

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, COCOanalyze_demo.ipynb, LICENSE, Makefile, README.md, analysisAPI, annotations, detections, latex, pycocotools, requirements.txt, run_analysis.py, setup.py

Authors Matteo Ruggero Ronchi, Pietro Perona arXiv ID 1707.05388 Category cs.CV: Computer Vision Citations 180 Venue IEEE International Conference on Computer Vision Repository https://github.com/matteorr/coco-analyze โญ 233 Last Checked 1 month ago
Abstract
We propose a new method to analyze the impact of errors in algorithms for multi-instance pose estimation and a principled benchmark that can be used to compare them. We define and characterize three classes of errors - localization, scoring, and background - study how they are influenced by instance attributes and their impact on an algorithm's performance. Our technique is applied to compare the two leading methods for human pose estimation on the COCO Dataset, measure the sensitivity of pose estimation with respect to instance size, type and number of visible keypoints, clutter due to multiple instances, and the relative score of instances. The performance of algorithms, and the types of error they make, are highly dependent on all these variables, but mostly on the number of keypoints and the clutter. The analysis and software tools we propose offer a novel and insightful approach for understanding the behavior of pose estimation algorithms and an effective method for measuring their strengths and weaknesses.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision