Cargando…

Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition

Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we...

Descripción completa

Detalles Bibliográficos
Autores principales: Wei, Guixiang, Zhou, Huijian, Zhang, Liping, Wang, Jianji
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10221027/
https://www.ncbi.nlm.nih.gov/pubmed/37430654
http://dx.doi.org/10.3390/s23104741
_version_ 1785049358377418752
author Wei, Guixiang
Zhou, Huijian
Zhang, Liping
Wang, Jianji
author_facet Wei, Guixiang
Zhou, Huijian
Zhang, Liping
Wang, Jianji
author_sort Wei, Guixiang
collection PubMed
description Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we propose spatial–temporal self-attention enhanced graph convolutional networks (STSAE-GCNs) that can analyze RGB yoga video data captured by cameras or smartphones. In the STSAE-GCN, we build a spatial–temporal self-attention module (STSAM), which can effectively enhance the spatial–temporal expression ability of the model and improve the performance of the proposed model. The STSAM has the characteristics of plug-and-play so that it can be applied in other skeleton-based action recognition methods and improve their performance. To prove the effectiveness of the proposed model in recognizing fitness yoga actions, we collected 960 fitness yoga action video clips in 10 action classes and built the dataset Yoga10. The recognition accuracy of the model on Yoga10 achieves 93.83%, outperforming the state-of-the-art methods, which proves that this model can better recognize fitness yoga actions and help students learn fitness yoga independently.
format Online
Article
Text
id pubmed-10221027
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102210272023-05-28 Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition Wei, Guixiang Zhou, Huijian Zhang, Liping Wang, Jianji Sensors (Basel) Article Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we propose spatial–temporal self-attention enhanced graph convolutional networks (STSAE-GCNs) that can analyze RGB yoga video data captured by cameras or smartphones. In the STSAE-GCN, we build a spatial–temporal self-attention module (STSAM), which can effectively enhance the spatial–temporal expression ability of the model and improve the performance of the proposed model. The STSAM has the characteristics of plug-and-play so that it can be applied in other skeleton-based action recognition methods and improve their performance. To prove the effectiveness of the proposed model in recognizing fitness yoga actions, we collected 960 fitness yoga action video clips in 10 action classes and built the dataset Yoga10. The recognition accuracy of the model on Yoga10 achieves 93.83%, outperforming the state-of-the-art methods, which proves that this model can better recognize fitness yoga actions and help students learn fitness yoga independently. MDPI 2023-05-14 /pmc/articles/PMC10221027/ /pubmed/37430654 http://dx.doi.org/10.3390/s23104741 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wei, Guixiang
Zhou, Huijian
Zhang, Liping
Wang, Jianji
Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title_full Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title_fullStr Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title_full_unstemmed Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title_short Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
title_sort spatial–temporal self-attention enhanced graph convolutional networks for fitness yoga action recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10221027/
https://www.ncbi.nlm.nih.gov/pubmed/37430654
http://dx.doi.org/10.3390/s23104741
work_keys_str_mv AT weiguixiang spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition
AT zhouhuijian spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition
AT zhangliping spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition
AT wangjianji spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition