Cargando…

Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset

The development of action recognition models has shown great performance on various video datasets. Nevertheless, because there is no rich data on target actions in existing datasets, it is insufficient to perform action recognition applications required by industries. To satisfy this requirement, d...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Doyoung, Lee, Inwoong, Kim, Dohyung, Lee, Sanghoon
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8539691/
https://www.ncbi.nlm.nih.gov/pubmed/34695988
http://dx.doi.org/10.3390/s21206774
_version_ 1784588808461746176
author Kim, Doyoung
Lee, Inwoong
Kim, Dohyung
Lee, Sanghoon
author_facet Kim, Doyoung
Lee, Inwoong
Kim, Dohyung
Lee, Sanghoon
author_sort Kim, Doyoung
collection PubMed
description The development of action recognition models has shown great performance on various video datasets. Nevertheless, because there is no rich data on target actions in existing datasets, it is insufficient to perform action recognition applications required by industries. To satisfy this requirement, datasets composed of target actions with high availability have been created, but it is difficult to capture various characteristics in actual environments because video data are generated in a specific environment. In this paper, we introduce a new ETRI-Activity3D-LivingLab dataset, which provides action sequences in actual environments and helps to handle a network generalization issue due to the dataset shift. When the action recognition model is trained on the ETRI-Activity3D and KIST SynADL datasets and evaluated on the ETRI-Activity3D-LivingLab dataset, the performance can be severely degraded because the datasets were captured in different environments domains. To reduce this dataset shift between training and testing datasets, we propose a close-up of maximum activation, which magnifies the most activated part of a video input in detail. In addition, we present various experimental results and analysis that show the dataset shift and demonstrate the effectiveness of the proposed method.
format Online
Article
Text
id pubmed-8539691
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-85396912021-10-24 Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset Kim, Doyoung Lee, Inwoong Kim, Dohyung Lee, Sanghoon Sensors (Basel) Article The development of action recognition models has shown great performance on various video datasets. Nevertheless, because there is no rich data on target actions in existing datasets, it is insufficient to perform action recognition applications required by industries. To satisfy this requirement, datasets composed of target actions with high availability have been created, but it is difficult to capture various characteristics in actual environments because video data are generated in a specific environment. In this paper, we introduce a new ETRI-Activity3D-LivingLab dataset, which provides action sequences in actual environments and helps to handle a network generalization issue due to the dataset shift. When the action recognition model is trained on the ETRI-Activity3D and KIST SynADL datasets and evaluated on the ETRI-Activity3D-LivingLab dataset, the performance can be severely degraded because the datasets were captured in different environments domains. To reduce this dataset shift between training and testing datasets, we propose a close-up of maximum activation, which magnifies the most activated part of a video input in detail. In addition, we present various experimental results and analysis that show the dataset shift and demonstrate the effectiveness of the proposed method. MDPI 2021-10-12 /pmc/articles/PMC8539691/ /pubmed/34695988 http://dx.doi.org/10.3390/s21206774 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kim, Doyoung
Lee, Inwoong
Kim, Dohyung
Lee, Sanghoon
Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title_full Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title_fullStr Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title_full_unstemmed Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title_short Action Recognition Using Close-Up of Maximum Activation and ETRI-Activity3D LivingLab Dataset
title_sort action recognition using close-up of maximum activation and etri-activity3d livinglab dataset
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8539691/
https://www.ncbi.nlm.nih.gov/pubmed/34695988
http://dx.doi.org/10.3390/s21206774
work_keys_str_mv AT kimdoyoung actionrecognitionusingcloseupofmaximumactivationandetriactivity3dlivinglabdataset
AT leeinwoong actionrecognitionusingcloseupofmaximumactivationandetriactivity3dlivinglabdataset
AT kimdohyung actionrecognitionusingcloseupofmaximumactivationandetriactivity3dlivinglabdataset
AT leesanghoon actionrecognitionusingcloseupofmaximumactivationandetriactivity3dlivinglabdataset