Idempotent Unsupervised Representation Learning for Skeleton-Based Action Recognition
Generative models, as a powerful technique for generation, also gradually become a critical tool for recognition tasks. However, in skeleton-based action recognition, the features obtained from existing pre-trained generative methods contain redundant information unrelated to recognition, which cont...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Generative models, as a powerful technique for generation, also gradually
become a critical tool for recognition tasks. However, in skeleton-based action
recognition, the features obtained from existing pre-trained generative methods
contain redundant information unrelated to recognition, which contradicts the
nature of the skeleton's spatially sparse and temporally consistent properties,
leading to undesirable performance. To address this challenge, we make efforts
to bridge the gap in theory and methodology and propose a novel skeleton-based
idempotent generative model (IGM) for unsupervised representation learning.
More specifically, we first theoretically demonstrate the equivalence between
generative models and maximum entropy coding, which demonstrates a potential
route that makes the features of generative models more compact by introducing
contrastive learning. To this end, we introduce the idempotency constraint to
form a stronger consistency regularization in the feature space, to push the
features only to maintain the critical information of motion semantics for the
recognition task. Our extensive experiments on benchmark datasets, NTU RGB+D
and PKUMMD, demonstrate the effectiveness of our proposed method. On the NTU 60
xsub dataset, we observe a performance improvement from 84.6$\%$ to 86.2$\%$.
Furthermore, in zero-shot adaptation scenarios, our model demonstrates
significant efficacy by achieving promising results in cases that were
previously unrecognizable. Our project is available at
\url{https://github.com/LanglandsLin/IGM}. |
---|---|
DOI: | 10.48550/arxiv.2410.20349 |