Cargando…
Deep Learning-Based Real-Time Multiple-Person Action Recognition System
Action recognition has gained great attention in automatic video analysis, greatly reducing the cost of human resources for smart surveillance. Most methods, however, focus on the detection of only one action event for a single person in a well-segmented video, rather than the recognition of multipl...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506925/ https://www.ncbi.nlm.nih.gov/pubmed/32842485 http://dx.doi.org/10.3390/s20174758 |
_version_ | 1783585124621746176 |
---|---|
author | Tsai, Jen-Kai Hsu, Chen-Chien Wang, Wei-Yen Huang, Shao-Kang |
author_facet | Tsai, Jen-Kai Hsu, Chen-Chien Wang, Wei-Yen Huang, Shao-Kang |
author_sort | Tsai, Jen-Kai |
collection | PubMed |
description | Action recognition has gained great attention in automatic video analysis, greatly reducing the cost of human resources for smart surveillance. Most methods, however, focus on the detection of only one action event for a single person in a well-segmented video, rather than the recognition of multiple actions performed by more than one person at the same time for an untrimmed video. In this paper, we propose a deep learning-based multiple-person action recognition system for use in various real-time smart surveillance applications. By capturing a video stream of the scene, the proposed system can detect and track multiple people appearing in the scene and subsequently recognize their actions. Thanks to high resolution of the video frames, we establish a zoom-in function to obtain more satisfactory action recognition results when people in the scene become too far from the camera. To further improve the accuracy, recognition results from inflated 3D ConvNet (I3D) with multiple sliding windows are processed by a nonmaximum suppression (NMS) approach to obtain a more robust decision. Experimental results show that the proposed method can perform multiple-person action recognition in real time suitable for applications such as long-term care environments. |
format | Online Article Text |
id | pubmed-7506925 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75069252020-09-30 Deep Learning-Based Real-Time Multiple-Person Action Recognition System Tsai, Jen-Kai Hsu, Chen-Chien Wang, Wei-Yen Huang, Shao-Kang Sensors (Basel) Letter Action recognition has gained great attention in automatic video analysis, greatly reducing the cost of human resources for smart surveillance. Most methods, however, focus on the detection of only one action event for a single person in a well-segmented video, rather than the recognition of multiple actions performed by more than one person at the same time for an untrimmed video. In this paper, we propose a deep learning-based multiple-person action recognition system for use in various real-time smart surveillance applications. By capturing a video stream of the scene, the proposed system can detect and track multiple people appearing in the scene and subsequently recognize their actions. Thanks to high resolution of the video frames, we establish a zoom-in function to obtain more satisfactory action recognition results when people in the scene become too far from the camera. To further improve the accuracy, recognition results from inflated 3D ConvNet (I3D) with multiple sliding windows are processed by a nonmaximum suppression (NMS) approach to obtain a more robust decision. Experimental results show that the proposed method can perform multiple-person action recognition in real time suitable for applications such as long-term care environments. MDPI 2020-08-23 /pmc/articles/PMC7506925/ /pubmed/32842485 http://dx.doi.org/10.3390/s20174758 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Letter Tsai, Jen-Kai Hsu, Chen-Chien Wang, Wei-Yen Huang, Shao-Kang Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title | Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title_full | Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title_fullStr | Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title_full_unstemmed | Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title_short | Deep Learning-Based Real-Time Multiple-Person Action Recognition System |
title_sort | deep learning-based real-time multiple-person action recognition system |
topic | Letter |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506925/ https://www.ncbi.nlm.nih.gov/pubmed/32842485 http://dx.doi.org/10.3390/s20174758 |
work_keys_str_mv | AT tsaijenkai deeplearningbasedrealtimemultiplepersonactionrecognitionsystem AT hsuchenchien deeplearningbasedrealtimemultiplepersonactionrecognitionsystem AT wangweiyen deeplearningbasedrealtimemultiplepersonactionrecognitionsystem AT huangshaokang deeplearningbasedrealtimemultiplepersonactionrecognitionsystem |