Cargando…
Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation
Contrastive learning has received increasing attention in the field of skeleton-based action representations in recent years. Most contrastive learning methods use simple augmentation strategies to construct pairs of positive samples. When using such pairs of positive samples to learn action represe...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9698072/ https://www.ncbi.nlm.nih.gov/pubmed/36433585 http://dx.doi.org/10.3390/s22228989 |
_version_ | 1784838724867063808 |
---|---|
author | Zhou, Hualing Li, Xi Xu, Dahong Liu, Hong Guo, Jianping Zhang, Yihan |
author_facet | Zhou, Hualing Li, Xi Xu, Dahong Liu, Hong Guo, Jianping Zhang, Yihan |
author_sort | Zhou, Hualing |
collection | PubMed |
description | Contrastive learning has received increasing attention in the field of skeleton-based action representations in recent years. Most contrastive learning methods use simple augmentation strategies to construct pairs of positive samples. When using such pairs of positive samples to learn action representations, deeper feature information cannot be learned, thus affecting the performance of downstream tasks. To solve the problem of insufficient learning ability, we propose an asymmetric data augmentation strategy and attempt to apply it to the training of 3D skeleton-based action representations. First, we carefully study the different characteristics presented by different skeleton views and choose a specific augmentation method for a certain view. Second, specific augmentation methods are incorporated into the left and right branches of the asymmetric data augmentation pipeline to increase the convergence difficulty of the contrastive learning task, thereby significantly improving the quality of the learned action representations. Finally, since many methods directly act on the joint view, the augmented samples are quite different from the original samples. We use random probability activation to transform the joint view to avoid extreme augmentation of the joint view. Extensive experiments on NTU RGB + D datasets show that our method is effective. |
format | Online Article Text |
id | pubmed-9698072 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-96980722022-11-26 Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation Zhou, Hualing Li, Xi Xu, Dahong Liu, Hong Guo, Jianping Zhang, Yihan Sensors (Basel) Article Contrastive learning has received increasing attention in the field of skeleton-based action representations in recent years. Most contrastive learning methods use simple augmentation strategies to construct pairs of positive samples. When using such pairs of positive samples to learn action representations, deeper feature information cannot be learned, thus affecting the performance of downstream tasks. To solve the problem of insufficient learning ability, we propose an asymmetric data augmentation strategy and attempt to apply it to the training of 3D skeleton-based action representations. First, we carefully study the different characteristics presented by different skeleton views and choose a specific augmentation method for a certain view. Second, specific augmentation methods are incorporated into the left and right branches of the asymmetric data augmentation pipeline to increase the convergence difficulty of the contrastive learning task, thereby significantly improving the quality of the learned action representations. Finally, since many methods directly act on the joint view, the augmented samples are quite different from the original samples. We use random probability activation to transform the joint view to avoid extreme augmentation of the joint view. Extensive experiments on NTU RGB + D datasets show that our method is effective. MDPI 2022-11-20 /pmc/articles/PMC9698072/ /pubmed/36433585 http://dx.doi.org/10.3390/s22228989 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhou, Hualing Li, Xi Xu, Dahong Liu, Hong Guo, Jianping Zhang, Yihan Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title | Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title_full | Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title_fullStr | Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title_full_unstemmed | Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title_short | Self-Supervised Action Representation Learning Based on Asymmetric Skeleton Data Augmentation |
title_sort | self-supervised action representation learning based on asymmetric skeleton data augmentation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9698072/ https://www.ncbi.nlm.nih.gov/pubmed/36433585 http://dx.doi.org/10.3390/s22228989 |
work_keys_str_mv | AT zhouhualing selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation AT lixi selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation AT xudahong selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation AT liuhong selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation AT guojianping selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation AT zhangyihan selfsupervisedactionrepresentationlearningbasedonasymmetricskeletondataaugmentation |