Exploring Self-supervised Skeleton-based Action Recognition in Occluded Environments

September 21, 2023 ยท Entered Twilight ยท ๐Ÿ› IEEE International Joint Conference on Neural Network

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: OPSTL.code-workspace, README.md, assets, config.py, dataset.py, generate_NTU60_occ, logger.py, model.py, module, procedure.py, skeleton_imputation, utils

Authors Yifei Chen, Kunyu Peng, Alina Roitberg, David Schneider, Jiaming Zhang, Junwei Zheng, Yufan Chen, Ruiping Liu, Kailun Yang, Rainer Stiefelhagen arXiv ID 2309.12029 Category cs.CV: Computer Vision Cross-listed cs.MM, cs.RO, eess.IV Citations 1 Venue IEEE International Joint Conference on Neural Network Repository https://github.com/cyfml/OPSTL โญ 14 Last Checked 1 month ago
Abstract
To integrate action recognition into autonomous robotic systems, it is essential to address challenges such as person occlusions-a common yet often overlooked scenario in existing self-supervised skeleton-based action recognition methods. In this work, we propose IosPSTL, a simple and effective self-supervised learning framework designed to handle occlusions. IosPSTL combines a cluster-agnostic KNN imputer with an Occluded Partial Spatio-Temporal Learning (OPSTL) strategy. First, we pre-train the model on occluded skeleton sequences. Then, we introduce a cluster-agnostic KNN imputer that performs semantic grouping using k-means clustering on sequence embeddings. It imputes missing skeleton data by applying K-Nearest Neighbors in the latent space, leveraging nearby sample representations to restore occluded joints. This imputation generates more complete skeleton sequences, which significantly benefits downstream self-supervised models. To further enhance learning, the OPSTL module incorporates Adaptive Spatial Masking (ASM) to make better use of intact, high-quality skeleton sequences during training. Our method achieves state-of-the-art performance on the occluded versions of the NTU-60 and NTU-120 datasets, demonstrating its robustness and effectiveness under challenging conditions. Code is available at https://github.com/cyfml/OPSTL.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision