Cargando…

Gaze-driven placement of items for proactive visual exploration

Recent advances in digital signage technology have improved the ability to visually select specific items within a group. Although this is due to the ability to dynamically update the display of items, the corresponding layout schemes remain a subject of research. This paper explores the sophisticat...

Descripción completa

Detalles Bibliográficos
Autores principales: Takahashi, Shigeo, Uchita, Akane, Watanabe, Kazuho, Arikawa, Masatoshi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer Berlin Heidelberg 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8581132/
https://www.ncbi.nlm.nih.gov/pubmed/34785979
http://dx.doi.org/10.1007/s12650-021-00808-5
_version_ 1784596738570452992
author Takahashi, Shigeo
Uchita, Akane
Watanabe, Kazuho
Arikawa, Masatoshi
author_facet Takahashi, Shigeo
Uchita, Akane
Watanabe, Kazuho
Arikawa, Masatoshi
author_sort Takahashi, Shigeo
collection PubMed
description Recent advances in digital signage technology have improved the ability to visually select specific items within a group. Although this is due to the ability to dynamically update the display of items, the corresponding layout schemes remain a subject of research. This paper explores the sophisticated layout of items by respecting the underlying context of searching for favorite items. Our study begins by formulating the static placement of items as an optimization problem that incorporates aesthetic layout criteria as constraints. This is further extended to accommodate the dynamic placement of items for more proactive visual exploration based on the ongoing search context. Our animated layout is driven by analyzing the distribution of eye gaze through an eye-tracking device, by which we infer how the most attractive items lead to the finally wanted ones. We create a planar layout of items as a context map to establish association rules to dynamically replace existing items with new ones. For this purpose, we extract the set of important topics from a set of annotated texts associated with the items using matrix factorization. We also conduct user studies to evaluate the validity of the design criteria incorporated into both static and dynamic placement of items. After discussing the pros and cons of the proposed approach and possible themes for future research, we conclude this paper.
format Online
Article
Text
id pubmed-8581132
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer Berlin Heidelberg
record_format MEDLINE/PubMed
spelling pubmed-85811322021-11-12 Gaze-driven placement of items for proactive visual exploration Takahashi, Shigeo Uchita, Akane Watanabe, Kazuho Arikawa, Masatoshi J Vis (Tokyo) Regular Paper Recent advances in digital signage technology have improved the ability to visually select specific items within a group. Although this is due to the ability to dynamically update the display of items, the corresponding layout schemes remain a subject of research. This paper explores the sophisticated layout of items by respecting the underlying context of searching for favorite items. Our study begins by formulating the static placement of items as an optimization problem that incorporates aesthetic layout criteria as constraints. This is further extended to accommodate the dynamic placement of items for more proactive visual exploration based on the ongoing search context. Our animated layout is driven by analyzing the distribution of eye gaze through an eye-tracking device, by which we infer how the most attractive items lead to the finally wanted ones. We create a planar layout of items as a context map to establish association rules to dynamically replace existing items with new ones. For this purpose, we extract the set of important topics from a set of annotated texts associated with the items using matrix factorization. We also conduct user studies to evaluate the validity of the design criteria incorporated into both static and dynamic placement of items. After discussing the pros and cons of the proposed approach and possible themes for future research, we conclude this paper. Springer Berlin Heidelberg 2021-11-11 2022 /pmc/articles/PMC8581132/ /pubmed/34785979 http://dx.doi.org/10.1007/s12650-021-00808-5 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Regular Paper
Takahashi, Shigeo
Uchita, Akane
Watanabe, Kazuho
Arikawa, Masatoshi
Gaze-driven placement of items for proactive visual exploration
title Gaze-driven placement of items for proactive visual exploration
title_full Gaze-driven placement of items for proactive visual exploration
title_fullStr Gaze-driven placement of items for proactive visual exploration
title_full_unstemmed Gaze-driven placement of items for proactive visual exploration
title_short Gaze-driven placement of items for proactive visual exploration
title_sort gaze-driven placement of items for proactive visual exploration
topic Regular Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8581132/
https://www.ncbi.nlm.nih.gov/pubmed/34785979
http://dx.doi.org/10.1007/s12650-021-00808-5
work_keys_str_mv AT takahashishigeo gazedrivenplacementofitemsforproactivevisualexploration
AT uchitaakane gazedrivenplacementofitemsforproactivevisualexploration
AT watanabekazuho gazedrivenplacementofitemsforproactivevisualexploration
AT arikawamasatoshi gazedrivenplacementofitemsforproactivevisualexploration