Cargando…

Emotional characteristic analysis of human gait while real-time movie viewing

Emotion recognition is useful in many applications such as preventing crime or improving customer satisfaction. Most of current methods are performed using facial features, which require close-up face information. Such information is difficult to capture with normal security cameras. The advantage o...

Descripción completa

Detalles Bibliográficos
Autores principales: Jianwattanapaisarn, Nitchan, Sumi, Kaoru, Utsumi, Akira, Khamsemanan, Nirattaya, Nattee, Cholwich
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9614342/
https://www.ncbi.nlm.nih.gov/pubmed/36311549
http://dx.doi.org/10.3389/frai.2022.989860
_version_ 1784820180174503936
author Jianwattanapaisarn, Nitchan
Sumi, Kaoru
Utsumi, Akira
Khamsemanan, Nirattaya
Nattee, Cholwich
author_facet Jianwattanapaisarn, Nitchan
Sumi, Kaoru
Utsumi, Akira
Khamsemanan, Nirattaya
Nattee, Cholwich
author_sort Jianwattanapaisarn, Nitchan
collection PubMed
description Emotion recognition is useful in many applications such as preventing crime or improving customer satisfaction. Most of current methods are performed using facial features, which require close-up face information. Such information is difficult to capture with normal security cameras. The advantage of using gait and posture over conventional biometrics such as facial features is that gaits and postures can be obtained unobtrusively from faraway, even in a noisy environment. This study aims to investigate and analyze the relationship between human emotions and their gaits or postures. We collected a dataset made from the input of 49 participants for our experiments. Subjects were instructed to walk naturally in a circular walking path, while watching emotion-inducing videos on Microsoft HoloLens 2 smart glasses. An OptiTrack motion-capturing system was used for recording the gaits and postures of participants. The angles between body parts and walking straightness were calculated as features for comparison of body-part movements while walking under different emotions. Results of statistical analyses show that the subjects' arm swings are significantly different among emotions. And the arm swings on one side of the body could reveal subjects' emotions more obviously than those on the other side. Our results suggest that the arm movements together with information of arm side and walking straightness can reveal the subjects' current emotions while walking. That is, emotions of humans are unconsciously expressed by their arm swings, especially by the left arm, when they are walking in a non-straight walking path. We found that arm swings in happy emotion are larger than arm swings in sad emotion. To the best of our knowledge, this study is the first to perform emotion induction by showing emotion-inducing videos to the participants using smart glasses during walking instead of showing videos before walking. This induction method is expected to be more consistent and more realistic than conventional methods. Our study will be useful for implementation of emotion recognition applications in real-world scenarios, since our emotion induction method and the walking direction we used are designed to mimic the real-time emotions of humans as they walk in a non-straight walking direction.
format Online
Article
Text
id pubmed-9614342
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-96143422022-10-29 Emotional characteristic analysis of human gait while real-time movie viewing Jianwattanapaisarn, Nitchan Sumi, Kaoru Utsumi, Akira Khamsemanan, Nirattaya Nattee, Cholwich Front Artif Intell Artificial Intelligence Emotion recognition is useful in many applications such as preventing crime or improving customer satisfaction. Most of current methods are performed using facial features, which require close-up face information. Such information is difficult to capture with normal security cameras. The advantage of using gait and posture over conventional biometrics such as facial features is that gaits and postures can be obtained unobtrusively from faraway, even in a noisy environment. This study aims to investigate and analyze the relationship between human emotions and their gaits or postures. We collected a dataset made from the input of 49 participants for our experiments. Subjects were instructed to walk naturally in a circular walking path, while watching emotion-inducing videos on Microsoft HoloLens 2 smart glasses. An OptiTrack motion-capturing system was used for recording the gaits and postures of participants. The angles between body parts and walking straightness were calculated as features for comparison of body-part movements while walking under different emotions. Results of statistical analyses show that the subjects' arm swings are significantly different among emotions. And the arm swings on one side of the body could reveal subjects' emotions more obviously than those on the other side. Our results suggest that the arm movements together with information of arm side and walking straightness can reveal the subjects' current emotions while walking. That is, emotions of humans are unconsciously expressed by their arm swings, especially by the left arm, when they are walking in a non-straight walking path. We found that arm swings in happy emotion are larger than arm swings in sad emotion. To the best of our knowledge, this study is the first to perform emotion induction by showing emotion-inducing videos to the participants using smart glasses during walking instead of showing videos before walking. This induction method is expected to be more consistent and more realistic than conventional methods. Our study will be useful for implementation of emotion recognition applications in real-world scenarios, since our emotion induction method and the walking direction we used are designed to mimic the real-time emotions of humans as they walk in a non-straight walking direction. Frontiers Media S.A. 2022-10-14 /pmc/articles/PMC9614342/ /pubmed/36311549 http://dx.doi.org/10.3389/frai.2022.989860 Text en Copyright © 2022 Jianwattanapaisarn, Sumi, Utsumi, Khamsemanan and Nattee. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Jianwattanapaisarn, Nitchan
Sumi, Kaoru
Utsumi, Akira
Khamsemanan, Nirattaya
Nattee, Cholwich
Emotional characteristic analysis of human gait while real-time movie viewing
title Emotional characteristic analysis of human gait while real-time movie viewing
title_full Emotional characteristic analysis of human gait while real-time movie viewing
title_fullStr Emotional characteristic analysis of human gait while real-time movie viewing
title_full_unstemmed Emotional characteristic analysis of human gait while real-time movie viewing
title_short Emotional characteristic analysis of human gait while real-time movie viewing
title_sort emotional characteristic analysis of human gait while real-time movie viewing
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9614342/
https://www.ncbi.nlm.nih.gov/pubmed/36311549
http://dx.doi.org/10.3389/frai.2022.989860
work_keys_str_mv AT jianwattanapaisarnnitchan emotionalcharacteristicanalysisofhumangaitwhilerealtimemovieviewing
AT sumikaoru emotionalcharacteristicanalysisofhumangaitwhilerealtimemovieviewing
AT utsumiakira emotionalcharacteristicanalysisofhumangaitwhilerealtimemovieviewing
AT khamsemanannirattaya emotionalcharacteristicanalysisofhumangaitwhilerealtimemovieviewing
AT natteecholwich emotionalcharacteristicanalysisofhumangaitwhilerealtimemovieviewing