Count What You Want: Exemplar Identification and Few-shot Counting of Human Actions in the Wild

December 28, 2023 ยท Entered Twilight ยท ๐Ÿ› AAAI Conference on Artificial Intelligence

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, .idea, Builder, Checkpoints, Config, Data, DataSyn, Dataset, Engine, Img, Models, Readme.MD, Register, Utils, eval.py, requirements.txt, train.py

Authors Yifeng Huang, Duc Duy Nguyen, Lam Nguyen, Cuong Pham, Minh Hoai arXiv ID 2312.17330 Category cs.CV: Computer Vision Cross-listed cs.AI Citations 3 Venue AAAI Conference on Artificial Intelligence Repository https://github.com/cvlab-stonybrook/ExRAC โญ 5 Last Checked 1 month ago
Abstract
This paper addresses the task of counting human actions of interest using sensor data from wearable devices. We propose a novel exemplar-based framework, allowing users to provide exemplars of the actions they want to count by vocalizing predefined sounds ''one'', ''two'', and ''three''. Our method first localizes temporal positions of these utterances from the audio sequence. These positions serve as the basis for identifying exemplars representing the action class of interest. A similarity map is then computed between the exemplars and the entire sensor data sequence, which is further fed into a density estimation module to generate a sequence of estimated density values. Summing these density values provides the final count. To develop and evaluate our approach, we introduce a diverse and realistic dataset consisting of real-world data from 37 subjects and 50 action categories, encompassing both sensor and audio data. The experiments on this dataset demonstrate the viability of the proposed method in counting instances of actions from new classes and subjects that were not part of the training data. On average, the discrepancy between the predicted count and the ground truth value is 7.47, significantly lower than the errors of the frequency-based and transformer-based methods. Our project, code and dataset can be found at https://github.com/cvlab-stonybrook/ExRAC.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision