Cargando…

3D Gaze Estimation Using RGB-IR Cameras

In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to conti...

Descripción completa

Detalles Bibliográficos
Autores principales: Mokatren, Moayad, Kuflik, Tsvi, Shimshoni, Ilan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823916/
https://www.ncbi.nlm.nih.gov/pubmed/36616978
http://dx.doi.org/10.3390/s23010381
_version_ 1784866280532082688
author Mokatren, Moayad
Kuflik, Tsvi
Shimshoni, Ilan
author_facet Mokatren, Moayad
Kuflik, Tsvi
Shimshoni, Ilan
author_sort Mokatren, Moayad
collection PubMed
description In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to continuously and reliably track the pupil and the RGB camera is used to acquire corneal images of the same eye. Deep learning algorithms are trained to detect the pupil in IR and RGB images and to compute a per user 3D model of the eye in real time. Once the 3D model is built, the 3D gaze direction is computed starting from the eyeball center and passing through the pupil center to the outside world. This model can also be used to transform the pupil position detected in the IR image into its corresponding position in the RGB image and to detect the gaze direction in the corneal image. This technique circumvents the problem of pupil detection in RGB images, which is especially difficult and unreliable when the scene is reflected in the corneal images. In our approach, the auto-calibration process is transparent and unobtrusive. Users do not have to be instructed to look at specific objects to calibrate the eye tracker. They need only to act and gaze normally. The framework was evaluated in a user study in realistic settings and the results are promising. It achieved a very low 3D gaze error (2.12°) and very high accuracy in acquiring corneal images (intersection over union— [Formula: see text] = 0.71). The framework may be used in a variety of real-world mobile scenarios (indoors, indoors near windows and outdoors) with high accuracy.
format Online
Article
Text
id pubmed-9823916
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-98239162023-01-08 3D Gaze Estimation Using RGB-IR Cameras Mokatren, Moayad Kuflik, Tsvi Shimshoni, Ilan Sensors (Basel) Article In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to continuously and reliably track the pupil and the RGB camera is used to acquire corneal images of the same eye. Deep learning algorithms are trained to detect the pupil in IR and RGB images and to compute a per user 3D model of the eye in real time. Once the 3D model is built, the 3D gaze direction is computed starting from the eyeball center and passing through the pupil center to the outside world. This model can also be used to transform the pupil position detected in the IR image into its corresponding position in the RGB image and to detect the gaze direction in the corneal image. This technique circumvents the problem of pupil detection in RGB images, which is especially difficult and unreliable when the scene is reflected in the corneal images. In our approach, the auto-calibration process is transparent and unobtrusive. Users do not have to be instructed to look at specific objects to calibrate the eye tracker. They need only to act and gaze normally. The framework was evaluated in a user study in realistic settings and the results are promising. It achieved a very low 3D gaze error (2.12°) and very high accuracy in acquiring corneal images (intersection over union— [Formula: see text] = 0.71). The framework may be used in a variety of real-world mobile scenarios (indoors, indoors near windows and outdoors) with high accuracy. MDPI 2022-12-29 /pmc/articles/PMC9823916/ /pubmed/36616978 http://dx.doi.org/10.3390/s23010381 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Mokatren, Moayad
Kuflik, Tsvi
Shimshoni, Ilan
3D Gaze Estimation Using RGB-IR Cameras
title 3D Gaze Estimation Using RGB-IR Cameras
title_full 3D Gaze Estimation Using RGB-IR Cameras
title_fullStr 3D Gaze Estimation Using RGB-IR Cameras
title_full_unstemmed 3D Gaze Estimation Using RGB-IR Cameras
title_short 3D Gaze Estimation Using RGB-IR Cameras
title_sort 3d gaze estimation using rgb-ir cameras
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823916/
https://www.ncbi.nlm.nih.gov/pubmed/36616978
http://dx.doi.org/10.3390/s23010381
work_keys_str_mv AT mokatrenmoayad 3dgazeestimationusingrgbircameras
AT kufliktsvi 3dgazeestimationusingrgbircameras
AT shimshoniilan 3dgazeestimationusingrgbircameras