Cargando…
Gaze Tracking Based on Concatenating Spatial-Temporal Features
Based on experimental observations, there is a correlation between time and consecutive gaze positions in visual behaviors. Previous studies on gaze point estimation usually use images as the input for model trainings without taking into account the sequence relationship between image data. In addit...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8781122/ https://www.ncbi.nlm.nih.gov/pubmed/35062502 http://dx.doi.org/10.3390/s22020545 |
_version_ | 1784638013527031808 |
---|---|
author | Hwang, Bor-Jiunn Chen, Hui-Hui Hsieh, Chaur-Heh Huang, Deng-Yu |
author_facet | Hwang, Bor-Jiunn Chen, Hui-Hui Hsieh, Chaur-Heh Huang, Deng-Yu |
author_sort | Hwang, Bor-Jiunn |
collection | PubMed |
description | Based on experimental observations, there is a correlation between time and consecutive gaze positions in visual behaviors. Previous studies on gaze point estimation usually use images as the input for model trainings without taking into account the sequence relationship between image data. In addition to the spatial features, the temporal features are considered to improve the accuracy in this paper by using videos instead of images as the input data. To be able to capture spatial and temporal features at the same time, the convolutional neural network (CNN) and long short-term memory (LSTM) network are introduced to build a training model. In this way, CNN is used to extract the spatial features, and LSTM correlates temporal features. This paper presents a CNN Concatenating LSTM network (CCLN) that concatenates spatial and temporal features to improve the performance of gaze estimation in the case of time-series videos as the input training data. In addition, the proposed model can be optimized by exploring the numbers of LSTM layers, the influence of batch normalization (BN) and global average pooling layer (GAP) on CCLN. It is generally believed that larger amounts of training data will lead to better models. To provide data for training and prediction, we propose a method for constructing datasets of video for gaze point estimation. The issues are studied, including the effectiveness of different commonly used general models and the impact of transfer learning. Through exhaustive evaluation, it has been proved that the proposed method achieves a better prediction accuracy than the existing CNN-based methods. Finally, 93.1% of the best model and 92.6% of the general model MobileNet are obtained. |
format | Online Article Text |
id | pubmed-8781122 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87811222022-01-22 Gaze Tracking Based on Concatenating Spatial-Temporal Features Hwang, Bor-Jiunn Chen, Hui-Hui Hsieh, Chaur-Heh Huang, Deng-Yu Sensors (Basel) Article Based on experimental observations, there is a correlation between time and consecutive gaze positions in visual behaviors. Previous studies on gaze point estimation usually use images as the input for model trainings without taking into account the sequence relationship between image data. In addition to the spatial features, the temporal features are considered to improve the accuracy in this paper by using videos instead of images as the input data. To be able to capture spatial and temporal features at the same time, the convolutional neural network (CNN) and long short-term memory (LSTM) network are introduced to build a training model. In this way, CNN is used to extract the spatial features, and LSTM correlates temporal features. This paper presents a CNN Concatenating LSTM network (CCLN) that concatenates spatial and temporal features to improve the performance of gaze estimation in the case of time-series videos as the input training data. In addition, the proposed model can be optimized by exploring the numbers of LSTM layers, the influence of batch normalization (BN) and global average pooling layer (GAP) on CCLN. It is generally believed that larger amounts of training data will lead to better models. To provide data for training and prediction, we propose a method for constructing datasets of video for gaze point estimation. The issues are studied, including the effectiveness of different commonly used general models and the impact of transfer learning. Through exhaustive evaluation, it has been proved that the proposed method achieves a better prediction accuracy than the existing CNN-based methods. Finally, 93.1% of the best model and 92.6% of the general model MobileNet are obtained. MDPI 2022-01-11 /pmc/articles/PMC8781122/ /pubmed/35062502 http://dx.doi.org/10.3390/s22020545 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Hwang, Bor-Jiunn Chen, Hui-Hui Hsieh, Chaur-Heh Huang, Deng-Yu Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title | Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title_full | Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title_fullStr | Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title_full_unstemmed | Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title_short | Gaze Tracking Based on Concatenating Spatial-Temporal Features |
title_sort | gaze tracking based on concatenating spatial-temporal features |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8781122/ https://www.ncbi.nlm.nih.gov/pubmed/35062502 http://dx.doi.org/10.3390/s22020545 |
work_keys_str_mv | AT hwangborjiunn gazetrackingbasedonconcatenatingspatialtemporalfeatures AT chenhuihui gazetrackingbasedonconcatenatingspatialtemporalfeatures AT hsiehchaurheh gazetrackingbasedonconcatenatingspatialtemporalfeatures AT huangdengyu gazetrackingbasedonconcatenatingspatialtemporalfeatures |