Cargando…

Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition

Spatiotemporal graph convolution has made significant progress in skeleton-based action recognition in recent years. Most of the existing graph convolution methods take all the joints of the human skeleton as the overall modeling graph, ignoring the differences in the movement patterns of various pa...

Descripción completa

Detalles Bibliográficos
Autores principales: Zuo, Qi, Zou, Lian, Fan, Cien, Li, Dongqian, Jiang, Hao, Liu, Yifeng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7763937/
https://www.ncbi.nlm.nih.gov/pubmed/33322231
http://dx.doi.org/10.3390/s20247149
_version_ 1783628137217654784
author Zuo, Qi
Zou, Lian
Fan, Cien
Li, Dongqian
Jiang, Hao
Liu, Yifeng
author_facet Zuo, Qi
Zou, Lian
Fan, Cien
Li, Dongqian
Jiang, Hao
Liu, Yifeng
author_sort Zuo, Qi
collection PubMed
description Spatiotemporal graph convolution has made significant progress in skeleton-based action recognition in recent years. Most of the existing graph convolution methods take all the joints of the human skeleton as the overall modeling graph, ignoring the differences in the movement patterns of various parts of the human, and cannot well connect the relationship between the different parts of the human skeleton. To capture the unique features of different parts of human skeleton data and the correlation of different parts, we propose two new graph convolution methods: the whole graph convolution network (WGCN) and the part graph convolution network (PGCN). WGCN learns the whole scale skeleton spatiotemporal features according to the movement patterns and physical structure of the human skeleton. PGCN divides the human skeleton graph into several subgraphs to learn the part scale spatiotemporal features. Moreover, we propose an adaptive fusion module that combines the two features for multiple complementary adaptive fusion to obtain more effective skeleton features. By coupling these proposals, we build a whole and part adaptive fusion graph convolution neural network (WPGCN) that outperforms previous state-of-the-art methods on three large-scale datasets: NTU RGB+D 60, NTU RGB+D 120, and Kinetics Skeleton 400.
format Online
Article
Text
id pubmed-7763937
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-77639372020-12-27 Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition Zuo, Qi Zou, Lian Fan, Cien Li, Dongqian Jiang, Hao Liu, Yifeng Sensors (Basel) Article Spatiotemporal graph convolution has made significant progress in skeleton-based action recognition in recent years. Most of the existing graph convolution methods take all the joints of the human skeleton as the overall modeling graph, ignoring the differences in the movement patterns of various parts of the human, and cannot well connect the relationship between the different parts of the human skeleton. To capture the unique features of different parts of human skeleton data and the correlation of different parts, we propose two new graph convolution methods: the whole graph convolution network (WGCN) and the part graph convolution network (PGCN). WGCN learns the whole scale skeleton spatiotemporal features according to the movement patterns and physical structure of the human skeleton. PGCN divides the human skeleton graph into several subgraphs to learn the part scale spatiotemporal features. Moreover, we propose an adaptive fusion module that combines the two features for multiple complementary adaptive fusion to obtain more effective skeleton features. By coupling these proposals, we build a whole and part adaptive fusion graph convolution neural network (WPGCN) that outperforms previous state-of-the-art methods on three large-scale datasets: NTU RGB+D 60, NTU RGB+D 120, and Kinetics Skeleton 400. MDPI 2020-12-13 /pmc/articles/PMC7763937/ /pubmed/33322231 http://dx.doi.org/10.3390/s20247149 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zuo, Qi
Zou, Lian
Fan, Cien
Li, Dongqian
Jiang, Hao
Liu, Yifeng
Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title_full Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title_fullStr Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title_full_unstemmed Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title_short Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition
title_sort whole and part adaptive fusion graph convolutional networks for skeleton-based action recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7763937/
https://www.ncbi.nlm.nih.gov/pubmed/33322231
http://dx.doi.org/10.3390/s20247149
work_keys_str_mv AT zuoqi wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition
AT zoulian wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition
AT fancien wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition
AT lidongqian wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition
AT jianghao wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition
AT liuyifeng wholeandpartadaptivefusiongraphconvolutionalnetworksforskeletonbasedactionrecognition