(eccv) Package eccv Warning: Package ‘hyperref’ is loaded with option ‘pagebackref’, which is *not* recommended for camera-ready version
11email: {linlilang, zjh2020, liujiaying}@pku.edu.cn
aladonwlh@pku.stu.edu.cn
Idempotent Unsupervised Representation Learning for Skeleton-Based Action Recognition
Abstract
Generative models, as a powerful technique for generation, also gradually become a critical tool for recognition tasks. However, in skeleton-based action recognition, the features obtained from existing pre-trained generative methods contain redundant information unrelated to recognition, which contradicts the nature of the skeleton’s spatially sparse and temporally consistent properties, leading to undesirable performance. To address this challenge, we make efforts to bridge the gap in theory and methodology and propose a novel skeleton-based idempotent generative model (IGM) for unsupervised representation learning. More specifically, we first theoretically demonstrate the equivalence between generative models and maximum entropy coding, which demonstrates a potential route that makes the features of generative models more compact by introducing contrastive learning. To this end, we introduce the idempotency constraint to form a stronger consistency regularization in the feature space, to push the features only to maintain the critical information of motion semantics for the recognition task. Our extensive experiments on benchmark datasets, NTU RGB+D and PKUMMD, demonstrate the effectiveness of our proposed method. On the NTU 60 xsub dataset, we observe a performance improvement from 84.6 to 86.2. Furthermore, in zero-shot adaptation scenarios, our model demonstrates significant efficacy by achieving promising results in cases that were previously unrecognizable. Our project is available at https://github.com/LanglandsLin/IGM.
Keywords:
Self-supervised learning skeleton-based action recognition contrastive learning1 Introduction
Skeletons represent human joints through 3D coordinate locations, providing a compact and efficient modality of representing human motion compared to RGB videos and depth data. Owing to their simplicity and superior discriminative capabilities for analysis, skeleton representations have been extensively employed in the field of action recognition tasks [68, 25, 66, 42, 32, 43].
Supervised skeleton-based action recognition methods [39, 37, 4] have demonstrated remarkable performance. However, they heavily rely on vast amounts of labeled training data, the collection of which can be a costly and time-consuming process. In order to reduce the reliance on fully supervised paradigms, self-supervised learning approaches have been explored in the context of skeleton-based action recognition [69, 20, 43, 46].
In the context of self-supervised pretraining paradigms, most methods can be broadly classified into two categories: generative learning-based [43, 60, 18] and contrastive learning-based approaches. Generative learning-based methods typically model the spatial-temporal correlations by predicting or reconstructing the masked skeleton data. With long-term global motion dynamics, Zheng et al. [69] were the pioneers in introducing the concept of reconstructing masked skeleton data. The structure of Masked Auto-Encoder (MAE) was used by Mao et al. [28] to predict the velocity of the masked part thus obtaining motion information modelling. However, skeleton data is by nature spatially sparse and temporal consistent while MAE’s feature preserves too much appearance information, which will interfere with the recognition tasks.
On the other route, contrastive learning-based methods also have recently demonstrated remarkable potential. These methods utilize skeleton transformations to generate positive pairs and aim to maintain consistency in the embedding space. Rao et al. [34] introduced shearing and cropping as data augmentation techniques. Guo et al. [11] extended these efforts by suggesting additional augmentations, such as rotation, masking, and flipping, to further enhance the consistency of contrastive learning. Contrastive learning, aimed at high-level tasks like recognition, often requires data transformation to filter out task-irrelevant information. This process results in a significant loss of information in the extracted features and hampers the ability to capture fine-grained motion details.
However, previous research has typically focused on these two paradigms separately. Their ideas and technical advantages are complementary and can be augmented, which is still under-explored. To address this gap, we first seek theoretical inspiration about the relationship between generative models and contrastive learning. In detail, we find that generative methods are equivalent to maximum entropy coding. This fact naturally inspires building the generative models with the related idempotent constraint to form a novel idempotent generative model, which is exactly equivalent to spectral contrastive learning but with improved recognition capacities.
Building upon this theoretical foundation, we propose a novel idempotent generative model to promote consistency in the feature space. By enforcing idempotence at the feature and distribution levels, our model enriches features with semantic motion information, thereby reducing the domain gap and adapting the generative model better for recognition. Moreover, the features of generative models are span on principal components, which easily leads to dimensional collapse, as recognition tasks primarily rely on descriminative local details. To address this imbalance, we introduce an adapter that fuses encoder and generator features. This integration expands the effective feature dimension of the feature space, facilitating more robust and comprehensive representation. Our model attains outstanding results through self-supervised learning in comparison to contemporary state-of-the-art methods.
In summary, our contributions are three-folded:
-
•
We propose an idempotent generative model to combine the benefits of generative pre-training and contrastive learning, which is inspired by the theoretical fact of their intrinsic correlation. This cooperation makes the model focus on extracting more compact information related to motion semantics, and obtain more powerful high-level representation within the generative model framework.
-
•
We further propose to utilize a multiple idempotency feature constraint. Through feature and distribution idempotency constraints, the feature consistency is improved, leading to not only improved recognition capture but also the perceptual reconstruction quality of the generative model.
-
•
We employ an adapter to fuse the features from the high-level semantic encoder and low-level skeleton generator from different subspaces to expand the representation dimension. Experiments show that our module improves the effective dimension of the feature space and encodes rich information.
2 Related Work
2.1 Skeleton-Based Action Recognition
Skeleton sequences encode the motion trajectories of human joints, representing rich information about human actions. Thus, skeleton data serves as a suitable modality for human action recognition [68, 25, 66, 42]. Skeleton can be obtained by applying pose estimation algorithms on RGB videos or depth maps [38].
Early studies focused on extracting hand-designed spatial and temporal domain features from skeleton sequences for human movement recognition [49, 50, 51, 10, 27]. In later work, efforts were made to model the positional information and higher-order temporal difference information of the human skeleton [44, 49]. Additionally, graphical models were built by tracking the trajectory of human joints to capture joint information in video sequences [52].
Recently, there has been a surge of interest in using graph structures for learning models [52]. Graph Neural Network (GNN) is one such model capturing intra-graph dependencies through information transfer between nodes. Various approaches have been proposed, such as spatio-domain inference networks, recurrent neural networks (RNNs), and graph convolutional networks (GCNs), to exploit graph structures for human action recognition [40, 36, 59]. These models automatically learn spatio-temporal patterns from skeleton data, facilitating strong action generalization. Moreover, attention mechanisms, multiscale aggregation schemes, and lightweight convolution operations have been integrated into GCN-based models to enhance their effectiveness and reduce computational costs [25, 66, 6, 41].
2.2 Self-Supervised Learning
The self-supervised task aims to extract data features from a large amount of unlabelled data [8]. It can be widely used in semantic segmentation, image classification, action recognition and many other tasks [17, 31]. These tasks are mainly classified into methods based on reconstruction and on contrastive learning.
Reconstruction based approach after masking part of the original data, the network is used to reconstruct the masked part of the data. He et al. [14] proposed Mased Auto-Encoder (MAE) to encode the visible patches and decode the visible and masked patches. This approach has been extended to the video domain and has been used in several studies. These methods typically use a visual Transformer as the backbone network in order to perform the mask reconstruction task. Feichtenhofer et al. [9] extended the image-based masked auto-encoder to use spatio-temporal learning to randomly mask spatio-temporal segments of a video and learn am auto-encoder for reconstruction at the pixel-level reconstruction. Similarly, in MaskFeat, Wei et al. [55] used several video cubes and utilized the model to predict them using the remaining information.
Contrastive learning pushes pairs of positive sample together while pushing pairs of negative sample further apart. To generate negative samples, contrastive learning pairs anchor frames with frames from other videos. There are various ways of generating positive and negative samples, which is the main factor that distinguishes different contrastive methods.
Most of these methods generate positive and negative samples by different ways in order to minimize and maximise the distance between them respectively. In the image domain, positive samples are usually generated by enhancing the image in different ways [47, 56, 1, 62, 16]. These enhancements include rotation, cropping, random greyscale and colour change [2]. Scaling these methods in video can be difficult because each video comparison increases the memory required, especially if multiple enhancements are used for multiple positive samples. Another challenge is incorporating the temporal domain into the enhancement. Some methods simply apply the same enhancement in the image to each frame [48]. Some methods include additional frame alignments that may be based on the temporal domain [26]. Finally, some methods rely on motion and optical flow maps as positive samples [33].
3 Idempotency Generation Network (IGN)
3.1 Self-Conditional Generative Models as Maximum Entropy Coding
Self-conditional generative modeling [14] is frequently employed as a pre-training task in self-supervised learning. It is generally structured as an auto-encoder. Formally, given the input skeleton data , the reconstruction loss is:
(1) |
where , is the encoder, and is data transformation. is the generator. is the distance. is the data distribution and is the feature distribution given . is the conditional entropy. In the context of MAE, this data transformation represents masked data during training. Conversely, in denoising auto-encoders, this transformation signifies adding noise to the input data.
In the context of mutual information, this loss function is equivalent to optimizing the mutual information between the extracted features and the input data . Based on the relationship between mutual information and entropy, we get
(2) |
Since the entropy of remains constant, decreasing the reconstruction loss is akin to increasing the mutual information. Conversely, as the features are deterministically derived from the data by an encoder , the entropy tends towards zero. Hence, maximizing the mutual information is equivalent to maximizing the entropy of the feature space .
Estimating the true distributions of the representation space is exceedingly challenging. Following works [63, 24], we leverage lossy data coding, a computationally feasible alternative, as a surrogate for the entropy of continuous random variables . This approach involves determining the minimal number of bits required to encode a set of samples subject to a distortion , as defined by the coding length function below [47, 72]:
(3) |
where is the upper bound of the expected decoding error between and the decoded . is the determinant of a matrix. is the dimension of the feature space. Utilizing the identity , we derive , where Tr denotes the trace of the matrix and , . Finally, we apply a Taylor series expansion to the logarithm of the matrix to obtain:
(4) |
because the features are projected into spherical space , . Hence, the first term does not contribute to the reconstruction learning. In essence, self-conditional generation primarily diminishes the inter-data similarity within the feature space:
(5) |
where .
3.2 Idempotent Generative Models as Spectral Contrastive Learning
The idempotence of a self-conditional generative model refers to its stability in re-encoding [57]. More precisely, if we denote the original data as , the encoder as , the encoded feature as , the decoder as , and the reconstruction as , then the self-conditional generative model is considered idempotent:
(6) |
Idempotence is frequently employed in the generative domain to augment the perceptual loss of generated images. The idempotent loss is formulated as:
(7) |
where because we normalize the feature space. Therefore, the idempotent generative model maximizes the entropy of the feature space while simultaneously minimizing the feature distance between the data and the generated data. The total loss of the idempotent generative model is expressed as:
(8) | ||||
where is the adjacency matrix defined by the data generation and is a constant. The weights . is the Laplacian matrix. This demonstrates its equivalence to spectral contrastive learning. And the advantage of our approach over spectral contrastive learning is that we additionally optimise the residual term to capture higher order information.
Further, we exploit data idempotence and feature idempotence to enhance representation learning and action generation. The unified generative-perceptual model contains both an encoder and generator . And our idempotency constraints pay attention to both the data and the feature distributions, which improves both generation and feature learning.
(9) |
3.3 Relationship to Masked Auto-Encoder
As mentioned in Eq. 2, the generative network without idempotent constraints has a conditional entropy of 0 because the encoding process is deterministic. Idempotent generative networks, on the other hand, treat features as a random variable sampled from the distribution of features across all of the generated data for the same data , thus transforming into a non-deterministic process:
(10) |
where is the generation process. Therefore, idempotent constraints are essentially about diminishing conditional entropy , which in turn maximizes the mutual information between features and data.
In contrast, methods like MAE implicitly prioritize maximizing feature similarity across masked samples of the same data:
(11) |
where is the random masking process. Consequently, features from two distinct data that undergo similar transformed or generated data are clustered into the same class. However, the data obtained through data transformation may not be the real data and thus far from the real data distribution.
3.4 Relationship to Downstream Tasks
Through the analysis of previous work [13, 67, 54, 12, 7] on spectral contrastive learning, the error rate of the downstream linear evaluation can be bounded by the generated adjacency matrix and clustering error probabilities :
Theorem 3.1
If are the eigenvalues of , and if the clustering purity is , we obtain:
(12) |
where are some constants.
This theorem illustrates the constraints on accuracy imposed by the purity . A large purity and a small number of clusters result in a low error rate. When the diversity in the generated data is insufficient, the sum of small singular values of the adjacency matrix become large, resulting in less tightly clustered groups. Conversely, excessive diversity in the generated data may compromise the preservation of motion information, thereby increasing the error rate in clustering and undermining overall clustering effectiveness.
Therefore, to make the feature space of the idempotent generative model more capable of clustering, it is necessary to increase the diversity of the generated data for a stronger feature consistency constraint. However, a paradox is demonstrated here. Ordinary generative processes result in limited diversity under self-conditional generation due to constraints on the distance between the generated data and the original data. So in order to simultaneously obtain diverse and motion semantics preserving generated data, we propose an idempotent self-conditional generation model based on the diffusion generation model. The diversity of the generated data is provided by the noise sampling process of the diffusion model.

3.5 Idempotent Diffusion Generation Model
Our model consists of three parts, an encoder , a generator and an adapter . The encoder extracts features as conditions for the generator and also as inputs to the downstream task classifier . And the generator reconstructs the skeleton data based on the features. The adapter , in turn, is responsible for projecting and fusing the features extracted by the encoder into the generator’s feature space to be used as conditions.
Encoder : We start by applying some data augmentations to the data to obtain data for increasing diversity. Then, spatio-temporal position embeddings and are added after projection to the feature space by linear projection:
(13) |
Following that, layers of vanilla transformer blocks are employed to extract latent representations . Each block consists of a multi-head self-attention (MSA) module and a feed-forward network (FFN) module. Residual connections are utilized within each module, which are then followed by layer normalization (LN).
Generator : The generator and encoder maintain the same structure. But the input is the noise data obtained by sampling in the diffusion. The generator predicts the noise magnitude by taking noise data and feature conditions as inputs.
Adapter : The adapter merges the features extracted by the encoder into the generator . This is necessary because high-level tasks like recognition operate in a different feature space compared to low-level tasks like generation. Recognition tasks necessitate capturing high-frequency action movements while generation primarily focuses on optimizing principal component space (with large singular values), such as walking or waving, which rely more on bottom component subspace like velocity. Thus, we introduce a feature fusion method that decouples principal and bottom component subspace, allowing the encoder features to focus more on high-frequency information, making them more suitable for high-level tasks such as action recognition. These features are then injected into the bottom component feature space of the generator.
Manifold Decoupled Feature Fusion Module: To derive discriminative features for use as semantic guides, we draw inspiration from negative samples in contrastive learning. We assume that regions with motion semantics have the lowest similarity to other regions in the same sequence. , where is the length of tokens of . The uniformity loss in contrastive learning is:
(14) |
the derivative of is as follows:
(15) |
where and . . removes low-frequency information. Based on this analysis, we extract the high-frequency information of the features as semantic information:
(16) |
Through this high-pass filtering, we filter out some low-frequency information of principal component space such as the mean value in the sequences, which is not very meaningful for recognition, and retain the semantic information, which is more important for recognition. This module also mitigates dimensionality collapse, making features more informative.
We then fuse the features into the generator by replacing LayerNorm (LN) with Adaptive LayerNorm (AdaLN) with the following equation:
(17) |
where represents the hidden representation of the generator, and are obtained from linear projection of timestep embedding and high-frequency condition , respectively. Through AdaLN layers, the condition guides the denoising process by scaling and shifting the normalized hidden representation.
Idempotence Generation Loss: Our loss function comprises two components: the noise prediction loss of the diffusion model and the idempotency constraint.
Noise Prediction Loss: The diffusion model is trained by predicting the noise from the input noise data:
(18) | ||||
Idempotence Constraint: To obtain consistency constraints on features, we adopt two types of idempotency losses, feature idempotency constraint and distribution idempotency constraint.
1) Feature idempotency constraint performs on features. We use the predicted noise to perform a step of de-noising to get the estimated generated data :
(19) |
Therefore, the feature idempotency constraint based on this generated data is formulated as:
(20) | ||||
Since the generated data may be noisy, we input the features with noise and the number of time steps as auxiliary information.
2) Distribution idempotency constraint aims to align the feature distributions of the generated and original data. It is essential to maintain the manifold structure of the generated data consistent with the manifold of the original data. We capture the feature manifold structure through inter-feature similarity:
(21) |
where is -th token data. We align it to the feature structure of the original ground truth data:
(22) |
where is the distance metric between two distributions. The feature idempotency constraint captures richer structural information and allows for the construction of tighter clusters. This is because the adjacency matrix not only connects different generated data of the same data but also connects different data with similar features. Based on this idempotent alignment, we enhance the generative power of the model for stronger perceptual performance, while the encoder learns stronger feature consistency. This results in reduced singular values of the adjacency matrix and better downstream task performance.
4 Experiment Results
To evaluate the effectiveness of our approach, we conducted experiments on two benchmark datasets: the NTU RGB+D dataset [35, 23] and the PKUMMD dataset [22].
Models | Architecture | NTU 60 | NTU 120 | ||
---|---|---|---|---|---|
xview | xsub | xset | xsub | ||
Contrastive Learning: | |||||
3s-AimCLR [11] | GCN | 83.4 | 77.8 | 66.7 | 67.9 |
3s-CPM [64] | GCN | 84.9 | 78.7 | 69.6 | 68.7 |
3s-CMD [29] | GRU | 90.9 | 84.1 | 76.1 | 74.7 |
GL-Transformer [18] | Transformer | 83.8 | 76.3 | 68.7 | 66.0 |
3s-ActCLR [21] | GCN | 88.8 | 84.3 | 75.7 | 74.3 |
Generative Learning: | |||||
3s-Colorization [61] | DGCNN | 87.2 | 79.1 | 70.8 | 69.2 |
SkeletonMAE [58] | GCN | 77.7 | 74.8 | 73.5 | 72.5 |
MAMP [28] | Transformer | 89.1 | 84.9 | 79.1 | 78.6 |
Contrative Learning & Generative Learning: | |||||
CRRL [53] | GRU | 73.8 | 67.6 | 57.0 | 56.2 |
PCM3 [65] | GRU | 90.4 | 83.9 | 77.5 | 76.3 |
IGM (Ours) | Transformer | 91.2 | 86.2 | 81.4 | 80.0 |
Method | Joint Noise (, ) | Part Occlusion | ||
---|---|---|---|---|
(1.0, 0.1) | (1.0, 0.05) | (0.5, 0.1) | Right Arms | |
AimCLR [11] | 6.3 | 16.6 | 22.0 | 28.1 |
ActCLR [21] | 12.7 | 33.5 | 28.6 | 30.3 |
MAMP [28] | 2.4 | 6.1 | 5.8 | 10.7 |
IGM (Ours) | 58.7 | 63.0 | 65.3 | 56.9 |
4.1 Datasets and Settings
NTU RGB+D Dataset 60 (NTU 60) [35] comprises a comprehensive compilation of 56,578 videos, covering 60 unique action labels. Every video includes annotations detailing the positions of 25 joints for each body, illustrating interactions among pairs and individual activities.
NTU RGB+D Dataset 120 (NTU 120) [23] is one of the most comprehensive datasets for recognizing actions. Encompassing 114,480 videos, it spans 120 unique action categories. This dataset documents the performance of actions by 106 individuals across diverse environments, employing 32 distinct recording configurations.
PKU Multi-Modality Dataset (PKUMMD) [22] encompasses 52 action classes and nearly 20,000 instances, with each sample comprising 25 joints, thoroughly tackling the multi-modal 3D comprehension of human actions. The dataset is partitioned into two segments, with Part II showcasing more demanding data owing to heightened view diversity, resulting in skeleton noise.
For enhancing network training, all skeleton sequences undergo temporal downsampling to 120 frames. The encoder and generator are built using the Transformer architecture [59], employing hidden channels configured to a dimension of 256. To assess performance, we employ a fully connected layer .
To refine our network, we employ the Adam optimizer [30]. Training is executed on a single NVIDIA GeForce RTX 4090, employing a batch size of 128, and the network undergoes training for 400 epochs.
4.2 Evaluation and Comparison
For a comprehensive assessment, we conduct comparative analysis of our approach with other methodologies across diverse scenarios.
Linear Evaluation. In the linear evaluation framework, we utilize an encoder to process the extracted features and a linear classifier for action classification. The evaluation metric employed is the accuracy of action recognition. Notably, the encoder remains unchanged throughout the linear evaluation protocol. Our model demonstrates superior performance on the datasets outlined in Table 1 compared to other methodologies.
KNN Evaluation. In the K-Nearest Neighbors (KNN) evaluation setup, where the fixed encoder extracts features without any trainable parameters, our model showcases superiority in action recognition accuracy on the presented datasets. Table 2 highlights the effectiveness of our approach compared to other methods in this evaluation mechanism.
Transfer Learning. In the transfer learning scenario, we assess the generalization capability of our model by pretraining it on the source data using a self-supervised task. We then evaluate the model’s performance on the target dataset using the linear evaluation mechanism, with the encoder maintaining fixed parameters without additional fine-tuning. Our approach demonstrates superior performance in the transfer learning setting, as illustrated in Table 3.
Method | DDPM [15] | MDM [45] | SkeletonMAE [58] | IGM (Ours) |
---|---|---|---|---|
MPJPE (mm) | 130.2 | 87.6 | 329.7 | 79.2 |
FID | 1.78 | 1.26 | 2.69 | 1.18 |
Zero-Shot Domain Generalization. By applying 4 types of corruption to the validation sets of all datasets, we assess the generalization of our proposed method compared to baseline approaches. For joint noise, we add noise with a probability of with a variance of to some joints. We leverage the generative capability of our model, enabling us to denoise noisy skeleton data at test time. Subsequently, we utilize the generated skeleton data for recognition, significantly enhancing the generalization ability of our model. In Table 4, our proposed approach shows consistent and substantial performance improvements.
Reconstruction Evaluation. In this section, we implement IGM for mask prediction tasks. We input the masked data into the encoder to extract features as conditions for generation, noting that the reconstruction task does not require adding data transformations to the conditions. Our method is compared with diffusion-based methods DDPM and MDM in Table 5. Figs 3 and 4 show visualizations and feature visualizations of both the generated data and the ground truth data. Despite sharing the same feature distribution, the generated samples exhibit some diversity due to the noise introduced in the conditions.
Module | KNN | Linear | ||
---|---|---|---|---|
FFM | ||||
64.7 | 83.3 | |||
67.6 | 85.1 | |||
68.4 | 85.5 | |||
69.0 | 86.0 | |||
69.3 | 86.2 |




4.3 Ablation Study
Here’s the modified text for the ablation experiments:
Analysis of Module Combination. We investigate the performance of various combinations of modules and observe that each module contributes to a certain degree of improvement. Optimal performance is attained when all three modules are combined. As depicted in Table 6, each module enhances performance.
Analysis of Mitigating Dimensional Collapse. The analysis points out that the feature space of the generated model is susceptible to dimensionality collapse, resulting in the extracted features losing the information needed for recognition. Fig. 3 shows the feature space of the encoder trained using the generative model and the feature space after Adapter. The token after removing similarity by Adapter network has higher feature values, i.e., the dimension collapse phenomenon is mitigated.
5 Conclusions
In this research, we propose the skeleton-based idempotent generative model (IGM) for unsupervised representation learning, presenting a novel framework that maximizes the potential of generative models for representation learning. By implementing idempotence at both the feature level and distribution, our model enriches features with semantic information about motion, making them more suitable for recognition tasks. Additionally, as the generative model primarily focuses on the principal component space, it is more susceptible to dimensional collapse. Conversely, recognition tasks rely more on the bottom subspace. To address this imbalance, we design an adapter that fuses encoder and generator features from different subspaces, thereby enhancing the effective feature dimension of the feature space.
Acknowledgements
This work was supported in part by the National Natural Science Foundation of China under Grant No.62172020, and in part by the Key Laboratory of Science, Technology and Standard in Press Industry (Key Laboratory of Intelligent Press Media Technology).
References
- [1] Bachman, P., Hjelm, R.D., Buchwalter, W.: Learning representations by maximizing mutual information across views. In: Proc. Advances in Neural Information Processing Systems (2019)
- [2] Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: Proc. Int’l Conference for Machine Learning (2020)
- [3] Chen, Y., Zhao, L., Yuan, J., Tian, Y., Xia, Z., Geng, S., Han, L., Metaxas, D.N.: Hierarchically self-supervised transformer for human skeleton representation learning. In: Proc. European Conference on Computer Vision (2022)
- [4] Chen, Y., Zhang, Z., Yuan, C., Li, B., Deng, Y., Hu, W.: Channel-wise topology refinement graph convolution for skeleton-based action recognition. In: Proc. Int’l Conference on Computer Vision (2021)
- [5] Chen, Z., Liu, H., Guo, T., Chen, Z., Song, P., Tang, H.: Contrastive learning from spatio-temporal mixed skeleton sequences for self-supervised skeleton-based action recognition. arXiv:2207.03065 (2022)
- [6] Cheng, K., Zhang, Y., He, X., Chen, W., Cheng, J., Lu, H.: Skeleton-based action recognition with shift graph convolutional network. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. pp. 183–192 (2020)
- [7] Du, T., Wang, Y., Wang, Y.: On the role of discrete tokenization in visual representation learning. In: Proc. Int’l Conference on Learning Representations (2023)
- [8] Erhan, D., Bengio, Y., Courville, A., Manzagol, P.A., Vincent, P., Bengio, S.: Why does unsupervised pre-training help deep learning? Journal of Machine Learning Research 11(Feb), 625–660 (2010)
- [9] Feichtenhofer, C., Li, Y., He, K., et al.: Masked autoencoders as spatiotemporal learners. Proc. Advances in Neural Information Processing Systems 35, 35946–35958 (2022)
- [10] Goutsu, Y., Takano, W., Nakamura, Y.: Motion recognition employing multiple kernel learning of fisher vectors using local skeleton features. In: Proc. Int’l Conference for Machine Learning Workshops (2015)
- [11] Guo, T., Liu, H., Chen, Z., Liu, M., Wang, T., Ding, R.: Contrastive learning from extremely augmented skeleton sequences for self-supervised action recognition. Proc. AAAI Conference on Artificial Intelligence (2022)
- [12] Guo, X., Wang, Y., Du, T., Wang, Y.: Contranorm: A contrastive learning perspective on oversmoothing and beyond. arXiv preprint arXiv:2303.06562 (2023)
- [13] HaoChen, J., Wei, C., Gaidon, A., Ma, T.: Provable guarantees for self-supervised deep learning with spectral contrastive loss, 2021. arXiv preprint arXiv:2106.04156
- [14] He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2022)
- [15] Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Proc. Advances in Neural Information Processing Systems 33, 6840–6851 (2020)
- [16] Isola, P., Zoran, D., Krishnan, D., Adelson, E.H.: Learning visual groups from co-occurrences in space and time. arXiv:1511.06811 (2015)
- [17] Jang, E., Devin, C., Vanhoucke, V., Levine, S.: Grasp2Vec: Learning object representations from self-supervised grasping. arXiv:1811.06964 (2018)
- [18] Kim, B., Chang, H.J., Kim, J., Choi, J.Y.: Global-local motion transformer for unsupervised skeleton-based action learning. Proc. European Conference on Computer Vision (2022)
- [19] Li, L., Wang, M., Ni, B., Wang, H., Yang, J., Zhang, W.: 3D human action representation learning via cross-view consistency pursuit. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2021)
- [20] Lin, L., Song, S., Yang, W., Liu, J.: MS2L: Multi-task self-supervised learning for skeleton based action recognition. In: Proc. ACM Int’l Conference on Multimedia (2020)
- [21] Lin, L., Zhang, J., Liu, J.: Actionlet-dependent contrastive learning for unsupervised skeleton-based action recognition. In: CVPR (2023)
- [22] Liu, J., Song, S., Liu, C., Li, Y., Hu, Y.: A benchmark dataset and comparison study for multi-modal human action analytics. ACM Trans. on Multimedia Computing, Communications, and Applications (2020)
- [23] Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.Y., Kot, A.C.: NTU RGB+D 120: A large-scale benchmark for 3d human activity understanding. IEEE Trans. on Pattern Analysis and Machine Intelligence (2019)
- [24] Liu, X., Wang, Z., Li, Y.L., Wang, S.: Self-supervised learning via maximum entropy coding. Proc. Advances in Neural Information Processing Systems 35, 34091–34105 (2022)
- [25] Liu, Z., Zhang, H., Chen, Z., Wang, Z., Ouyang, W.: Disentangling and unifying graph convolutions for skeleton-based action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2020)
- [26] Lorre, G., Rabarisoa, J., Orcesi, A., Ainouz, S., Canu, S.: Temporal contrastive pretraining for video action recognition. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision. pp. 662–670 (2020)
- [27] Lv, F., Nevatia, R.: Recognition and segmentation of 3-d human action using hmm and multi-class adaboost. In: Proc. European Conference on Computer Vision (2006)
- [28] Mao, Y., Deng, J., Zhou, W., Fang, Y., Ouyang, W., Li, H.: Masked motion predictors are strong 3d action representation learners. In: Proc. Int’l Conference on Computer Vision. pp. 10181–10191 (2023)
- [29] Mao, Y., Zhou, W., Lu, Z., Deng, J., Li, H.: CMD: Self-supervised 3d action representation learning with cross-modal mutual distillation. Proc. European Conference on Computer Vision (2022)
- [30] Newey, W.K.: Adaptive estimation of regression models via moment restrictions. Journal of Econometrics (1988)
- [31] Owens, A., Efros, A.A.: Audio-visual scene analysis with self-supervised multisensory features. In: Proc. European Conference on Computer Vision (2018)
- [32] Peng, W., Hong, X., Chen, H., Zhao, G.: Learning graph convolutional network for skeleton-based human action recognition by neural searching. In: Proc. AAAI Conference on Artificial Intelligence (2020)
- [33] Rai, N., Adeli, E., Lee, K.H., Gaidon, A., Niebles, J.C.: Cocon: Cooperative-contrastive learning. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. pp. 3384–3393 (2021)
- [34] Rao, H., Xu, S., Hu, X., Cheng, J., Hu, B.: Augmented skeleton based contrastive action learning with momentum LSTM for unsupervised action recognition. Information Sciences (2021)
- [35] Shahroudy, A., Liu, J., Ng, T.T., Wang, G.: NTU RGB+D: A large scale dataset for 3d human activity analysis. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2016)
- [36] Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with directed graph neural networks. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. pp. 7912–7921 (2019)
- [37] Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2019)
- [38] Shotton, J., Sharp, T., Kipman, A., Fitzgibbon, A., Finocchio, M., Blake, A., Cook, M., Moore, R.: Real-time human pose recognition in parts from single depth images. Communications of the ACM (2013)
- [39] Si, C., Chen, W., Wang, W., Wang, L., Tan, T.: An attention enhanced graph convolutional lstm network for skeleton-based action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2019)
- [40] Si, C., Jing, Y., Wang, W., Wang, L., Tan, T.: Skeleton-based action recognition with spatial reasoning and temporal stack learning. In: Proc. European Conference on Computer Vision. pp. 103–118 (2018)
- [41] Song, Y.F., Zhang, Z., Shan, C., Wang, L.: Constructing stronger and faster baselines for skeleton-based action recognition. IEEE transactions on pattern analysis and machine intelligence 45(2), 1474–1488 (2022)
- [42] Song, Y., Zhang, Z., Shan, C., Wang, L.: Stronger, faster and more explainable: A graph convolutional baseline for skeleton-based action recognition. In: Proc. ACM Int’l Conference on Multimedia (2020)
- [43] Su, K., Liu, X., Shlizerman, E.: Predict & cluster: Unsupervised skeleton based action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2020)
- [44] Tao, L., Vidal, R.: Moving poselets: A discriminative and interpretable skeletal motion representation for action recognition. In: Proc. Int’l Conference for Machine Learning Workshops (2015)
- [45] Tevet, G., Raab, S., Gordon, B., Shafir, Y., Cohen-Or, D., Bermano, A.H.: Human motion diffusion model. arXiv preprint arXiv:2209.14916 (2022)
- [46] Thoker, F.M., Doughty, H., Snoek, C.G.: Skeleton-contrastive 3D action representation learning. In: Proc. ACM Int’l Conference on Multimedia (2021)
- [47] Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. arXiv:1906.05849 (2019)
- [48] Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. In: Proc. European Conference on Computer Vision. pp. 776–794. Springer (2020)
- [49] Vemulapalli, R., Arrate, F., Chellappa, R.: Human action recognition by representing 3d skeletons as points in a lie group. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2014)
- [50] Vemulapalli, R., Chellapa, R.: Rolling rotations for recognizing human actions from 3d skeletal data. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2016)
- [51] Wang, J., Liu, Z., Wu, Y., Yuan, J.: Mining actionlet ensemble for action recognition with depth cameras. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2012)
- [52] Wang, P., Yuan, C., Hu, W., Li, B., Zhang, Y.: Graph based skeleton motion representation and similarity measurement for action recognition. In: Proc. European Conference on Computer Vision. pp. 370–385 (2016)
- [53] Wang, P., Wen, J., Si, C., Qian, Y., Wang, L.: Contrast-reconstruction representation learning for self-supervised skeleton-based action recognition. IEEE Trans. on Image Processing 31, 6224–6238 (2022)
- [54] Wang, Y., Zhang, Q., Du, T., Yang, J., Lin, Z., Wang, Y.: A message passing perspective on learning dynamics of contrastive learning. arXiv preprint arXiv:2303.04435 (2023)
- [55] Wei, C., Fan, H., Xie, S., Wu, C.Y., Yuille, A., Feichtenhofer, C.: Masked feature prediction for self-supervised visual pre-training. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. pp. 14668–14678 (2022)
- [56] Wu, Z., Xiong, Y., Yu, S.X., Lin, D.: Unsupervised feature learning via non-parametric instance discrimination. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2018)
- [57] Xu, T., Zhu, Z., He, D., Li, Y., Guo, L., Wang, Y., Wang, Z., Qin, H., Wang, Y., Liu, J., et al.: Idempotence and perceptual image compression. arXiv preprint arXiv:2401.08920 (2024)
- [58] Yan, H., Liu, Y., Wei, Y., Li, Z., Li, G., Lin, L.: Skeletonmae: graph-based masked autoencoder for skeleton sequence pre-training. In: Proc. Int’l Conference on Computer Vision. pp. 5606–5618 (2023)
- [59] Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proc. AAAI Conference on Artificial Intelligence (2018)
- [60] Yang, S., Liu, J., Lu, S., Er, M.H., Kot, A.C.: Skeleton cloud colorization for unsupervised 3D action representation learning. In: Proc. Int’l Conference on Computer Vision (2021)
- [61] Yang, S., Liu, J., Lu, S., Hwa, E.M., Hu, Y., Kot, A.C.: Self-supervised 3d action representation learning with skeleton cloud colorization. IEEE Trans. on Pattern Analysis and Machine Intelligence (2023)
- [62] Ye, M., Zhang, X., Yuen, P.C., Chang, S.F.: Unsupervised embedding learning via invariant and spreading instance feature. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2019)
- [63] Yu, Y., Chan, K.H.R., You, C., Song, C., Ma, Y.: Learning diverse and discriminative representations via the principle of maximal coding rate reduction. Proc. Advances in Neural Information Processing Systems 33, 9422–9434 (2020)
- [64] Zhang, H., Hou, Y., Zhang, W., Li, W.: Contrastive positive mining for unsupervised 3d action representation learning. Proc. European Conference on Computer Vision (2022)
- [65] Zhang, J., Lin, L., Liu, J.: Prompted contrast with masked motion modeling: Towards versatile 3d action representation learning. In: Proc. ACM Int’l Conference on Multimedia. pp. 7175–7183 (2023)
- [66] Zhang, P., Lan, C., Zeng, W., Xing, J., Xue, J., Zheng, N.: Semantics-guided neural networks for efficient skeleton-based human action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2020)
- [67] Zhang, Q., Wang, Y., Wang, Y.: How mask matters: Towards theoretical understandings of masked autoencoders. Proc. Advances in Neural Information Processing Systems 35, 27127–27139 (2022)
- [68] Zhang, X., Xu, C., Tao, D.: Context aware graph convolution for skeleton-based action recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (2020)
- [69] Zheng, N., Wen, J., Liu, R., Long, L., Dai, J., Gong, Z.: Unsupervised representation learning with long-term dynamics for skeleton based action recognition. In: Proc. AAAI Conference on Artificial Intelligence (2018)