Cargando…

The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking

There is increasing evidence that hand gestures and speech synchronize their activity on multiple dimensions and timescales. For example, gesture’s kinematic peaks (e.g., maximum speed) are coupled with prosodic markers in speech. Such coupling operates on very short timescales at the level of sylla...

Descripción completa

Detalles Bibliográficos
Autores principales: Pouw, Wim, Trujillo, James P., Dixon, James A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148275/
https://www.ncbi.nlm.nih.gov/pubmed/31659689
http://dx.doi.org/10.3758/s13428-019-01271-9
_version_ 1783520559353561088
author Pouw, Wim
Trujillo, James P.
Dixon, James A.
author_facet Pouw, Wim
Trujillo, James P.
Dixon, James A.
author_sort Pouw, Wim
collection PubMed
description There is increasing evidence that hand gestures and speech synchronize their activity on multiple dimensions and timescales. For example, gesture’s kinematic peaks (e.g., maximum speed) are coupled with prosodic markers in speech. Such coupling operates on very short timescales at the level of syllables (200 ms), and therefore requires high-resolution measurement of gesture kinematics and speech acoustics. High-resolution speech analysis is common for gesture studies, given that field’s classic ties with (psycho)linguistics. However, the field has lagged behind in the objective study of gesture kinematics (e.g., as compared to research on instrumental action). Often kinematic peaks in gesture are measured by eye, where a “moment of maximum effort” is determined by several raters. In the present article, we provide a tutorial on more efficient methods to quantify the temporal properties of gesture kinematics, in which we focus on common challenges and possible solutions that come with the complexities of studying multimodal language. We further introduce and compare, using an actual gesture dataset (392 gesture events), the performance of two video-based motion-tracking methods (deep learning vs. pixel change) against a high-performance wired motion-tracking system (Polhemus Liberty). We show that the videography methods perform well in the temporal estimation of kinematic peaks, and thus provide a cheap alternative to expensive motion-tracking systems. We hope that the present article incites gesture researchers to embark on the widespread objective study of gesture kinematics and their relation to speech.
format Online
Article
Text
id pubmed-7148275
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-71482752020-04-16 The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking Pouw, Wim Trujillo, James P. Dixon, James A. Behav Res Methods Article There is increasing evidence that hand gestures and speech synchronize their activity on multiple dimensions and timescales. For example, gesture’s kinematic peaks (e.g., maximum speed) are coupled with prosodic markers in speech. Such coupling operates on very short timescales at the level of syllables (200 ms), and therefore requires high-resolution measurement of gesture kinematics and speech acoustics. High-resolution speech analysis is common for gesture studies, given that field’s classic ties with (psycho)linguistics. However, the field has lagged behind in the objective study of gesture kinematics (e.g., as compared to research on instrumental action). Often kinematic peaks in gesture are measured by eye, where a “moment of maximum effort” is determined by several raters. In the present article, we provide a tutorial on more efficient methods to quantify the temporal properties of gesture kinematics, in which we focus on common challenges and possible solutions that come with the complexities of studying multimodal language. We further introduce and compare, using an actual gesture dataset (392 gesture events), the performance of two video-based motion-tracking methods (deep learning vs. pixel change) against a high-performance wired motion-tracking system (Polhemus Liberty). We show that the videography methods perform well in the temporal estimation of kinematic peaks, and thus provide a cheap alternative to expensive motion-tracking systems. We hope that the present article incites gesture researchers to embark on the widespread objective study of gesture kinematics and their relation to speech. Springer US 2019-10-28 2020 /pmc/articles/PMC7148275/ /pubmed/31659689 http://dx.doi.org/10.3758/s13428-019-01271-9 Text en © The Author(s) 2019 Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
spellingShingle Article
Pouw, Wim
Trujillo, James P.
Dixon, James A.
The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title_full The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title_fullStr The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title_full_unstemmed The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title_short The quantification of gesture–speech synchrony: A tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
title_sort quantification of gesture–speech synchrony: a tutorial and validation of multimodal data acquisition using device-based and video-based motion tracking
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148275/
https://www.ncbi.nlm.nih.gov/pubmed/31659689
http://dx.doi.org/10.3758/s13428-019-01271-9
work_keys_str_mv AT pouwwim thequantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking
AT trujillojamesp thequantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking
AT dixonjamesa thequantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking
AT pouwwim quantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking
AT trujillojamesp quantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking
AT dixonjamesa quantificationofgesturespeechsynchronyatutorialandvalidationofmultimodaldataacquisitionusingdevicebasedandvideobasedmotiontracking