Cargando…
Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition
Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10221027/ https://www.ncbi.nlm.nih.gov/pubmed/37430654 http://dx.doi.org/10.3390/s23104741 |
_version_ | 1785049358377418752 |
---|---|
author | Wei, Guixiang Zhou, Huijian Zhang, Liping Wang, Jianji |
author_facet | Wei, Guixiang Zhou, Huijian Zhang, Liping Wang, Jianji |
author_sort | Wei, Guixiang |
collection | PubMed |
description | Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we propose spatial–temporal self-attention enhanced graph convolutional networks (STSAE-GCNs) that can analyze RGB yoga video data captured by cameras or smartphones. In the STSAE-GCN, we build a spatial–temporal self-attention module (STSAM), which can effectively enhance the spatial–temporal expression ability of the model and improve the performance of the proposed model. The STSAM has the characteristics of plug-and-play so that it can be applied in other skeleton-based action recognition methods and improve their performance. To prove the effectiveness of the proposed model in recognizing fitness yoga actions, we collected 960 fitness yoga action video clips in 10 action classes and built the dataset Yoga10. The recognition accuracy of the model on Yoga10 achieves 93.83%, outperforming the state-of-the-art methods, which proves that this model can better recognize fitness yoga actions and help students learn fitness yoga independently. |
format | Online Article Text |
id | pubmed-10221027 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-102210272023-05-28 Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition Wei, Guixiang Zhou, Huijian Zhang, Liping Wang, Jianji Sensors (Basel) Article Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we propose spatial–temporal self-attention enhanced graph convolutional networks (STSAE-GCNs) that can analyze RGB yoga video data captured by cameras or smartphones. In the STSAE-GCN, we build a spatial–temporal self-attention module (STSAM), which can effectively enhance the spatial–temporal expression ability of the model and improve the performance of the proposed model. The STSAM has the characteristics of plug-and-play so that it can be applied in other skeleton-based action recognition methods and improve their performance. To prove the effectiveness of the proposed model in recognizing fitness yoga actions, we collected 960 fitness yoga action video clips in 10 action classes and built the dataset Yoga10. The recognition accuracy of the model on Yoga10 achieves 93.83%, outperforming the state-of-the-art methods, which proves that this model can better recognize fitness yoga actions and help students learn fitness yoga independently. MDPI 2023-05-14 /pmc/articles/PMC10221027/ /pubmed/37430654 http://dx.doi.org/10.3390/s23104741 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wei, Guixiang Zhou, Huijian Zhang, Liping Wang, Jianji Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title | Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title_full | Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title_fullStr | Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title_full_unstemmed | Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title_short | Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition |
title_sort | spatial–temporal self-attention enhanced graph convolutional networks for fitness yoga action recognition |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10221027/ https://www.ncbi.nlm.nih.gov/pubmed/37430654 http://dx.doi.org/10.3390/s23104741 |
work_keys_str_mv | AT weiguixiang spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition AT zhouhuijian spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition AT zhangliping spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition AT wangjianji spatialtemporalselfattentionenhancedgraphconvolutionalnetworksforfitnessyogaactionrecognition |