Cargando…

What has been missed for predicting human attention in viewing driving clips?

Recent research progress on the topic of human visual attention allocation in scene perception and its simulation is based mainly on studies with static images. However, natural vision requires us to extract visual information that constantly changes due to egocentric movements or dynamics of the wo...

Descripción completa

Detalles Bibliográficos
Autores principales: Xu, Jiawei, Yue, Shigang, Menchinelli, Federica, Guo, Kun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5291110/
https://www.ncbi.nlm.nih.gov/pubmed/28168112
http://dx.doi.org/10.7717/peerj.2946
_version_ 1782504730799374336
author Xu, Jiawei
Yue, Shigang
Menchinelli, Federica
Guo, Kun
author_facet Xu, Jiawei
Yue, Shigang
Menchinelli, Federica
Guo, Kun
author_sort Xu, Jiawei
collection PubMed
description Recent research progress on the topic of human visual attention allocation in scene perception and its simulation is based mainly on studies with static images. However, natural vision requires us to extract visual information that constantly changes due to egocentric movements or dynamics of the world. It is unclear to what extent spatio-temporal regularity, an inherent regularity in dynamic vision, affects human gaze distribution and saliency computation in visual attention models. In this free-viewing eye-tracking study we manipulated the spatio-temporal regularity of traffic videos by presenting them in normal video sequence, reversed video sequence, normal frame sequence, and randomised frame sequence. The recorded human gaze allocation was then used as the ‘ground truth’ to examine the predictive ability of a number of state-of-the-art visual attention models. The analysis revealed high inter-observer agreement across individual human observers, but all the tested attention models performed significantly worse than humans. The inferior predictability of the models was evident from indistinguishable gaze prediction irrespective of stimuli presentation sequence, and weak central fixation bias. Our findings suggest that a realistic visual attention model for the processing of dynamic scenes should incorporate human visual sensitivity with spatio-temporal regularity and central fixation bias.
format Online
Article
Text
id pubmed-5291110
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-52911102017-02-06 What has been missed for predicting human attention in viewing driving clips? Xu, Jiawei Yue, Shigang Menchinelli, Federica Guo, Kun PeerJ Psychiatry and Psychology Recent research progress on the topic of human visual attention allocation in scene perception and its simulation is based mainly on studies with static images. However, natural vision requires us to extract visual information that constantly changes due to egocentric movements or dynamics of the world. It is unclear to what extent spatio-temporal regularity, an inherent regularity in dynamic vision, affects human gaze distribution and saliency computation in visual attention models. In this free-viewing eye-tracking study we manipulated the spatio-temporal regularity of traffic videos by presenting them in normal video sequence, reversed video sequence, normal frame sequence, and randomised frame sequence. The recorded human gaze allocation was then used as the ‘ground truth’ to examine the predictive ability of a number of state-of-the-art visual attention models. The analysis revealed high inter-observer agreement across individual human observers, but all the tested attention models performed significantly worse than humans. The inferior predictability of the models was evident from indistinguishable gaze prediction irrespective of stimuli presentation sequence, and weak central fixation bias. Our findings suggest that a realistic visual attention model for the processing of dynamic scenes should incorporate human visual sensitivity with spatio-temporal regularity and central fixation bias. PeerJ Inc. 2017-02-01 /pmc/articles/PMC5291110/ /pubmed/28168112 http://dx.doi.org/10.7717/peerj.2946 Text en ©2017 Xu et al. http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ) and either DOI or URL of the article must be cited.
spellingShingle Psychiatry and Psychology
Xu, Jiawei
Yue, Shigang
Menchinelli, Federica
Guo, Kun
What has been missed for predicting human attention in viewing driving clips?
title What has been missed for predicting human attention in viewing driving clips?
title_full What has been missed for predicting human attention in viewing driving clips?
title_fullStr What has been missed for predicting human attention in viewing driving clips?
title_full_unstemmed What has been missed for predicting human attention in viewing driving clips?
title_short What has been missed for predicting human attention in viewing driving clips?
title_sort what has been missed for predicting human attention in viewing driving clips?
topic Psychiatry and Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5291110/
https://www.ncbi.nlm.nih.gov/pubmed/28168112
http://dx.doi.org/10.7717/peerj.2946
work_keys_str_mv AT xujiawei whathasbeenmissedforpredictinghumanattentioninviewingdrivingclips
AT yueshigang whathasbeenmissedforpredictinghumanattentioninviewingdrivingclips
AT menchinellifederica whathasbeenmissedforpredictinghumanattentioninviewingdrivingclips
AT guokun whathasbeenmissedforpredictinghumanattentioninviewingdrivingclips