Cargando…
3D Gaze Estimation Using RGB-IR Cameras
In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to conti...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823916/ https://www.ncbi.nlm.nih.gov/pubmed/36616978 http://dx.doi.org/10.3390/s23010381 |
_version_ | 1784866280532082688 |
---|---|
author | Mokatren, Moayad Kuflik, Tsvi Shimshoni, Ilan |
author_facet | Mokatren, Moayad Kuflik, Tsvi Shimshoni, Ilan |
author_sort | Mokatren, Moayad |
collection | PubMed |
description | In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to continuously and reliably track the pupil and the RGB camera is used to acquire corneal images of the same eye. Deep learning algorithms are trained to detect the pupil in IR and RGB images and to compute a per user 3D model of the eye in real time. Once the 3D model is built, the 3D gaze direction is computed starting from the eyeball center and passing through the pupil center to the outside world. This model can also be used to transform the pupil position detected in the IR image into its corresponding position in the RGB image and to detect the gaze direction in the corneal image. This technique circumvents the problem of pupil detection in RGB images, which is especially difficult and unreliable when the scene is reflected in the corneal images. In our approach, the auto-calibration process is transparent and unobtrusive. Users do not have to be instructed to look at specific objects to calibrate the eye tracker. They need only to act and gaze normally. The framework was evaluated in a user study in realistic settings and the results are promising. It achieved a very low 3D gaze error (2.12°) and very high accuracy in acquiring corneal images (intersection over union— [Formula: see text] = 0.71). The framework may be used in a variety of real-world mobile scenarios (indoors, indoors near windows and outdoors) with high accuracy. |
format | Online Article Text |
id | pubmed-9823916 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-98239162023-01-08 3D Gaze Estimation Using RGB-IR Cameras Mokatren, Moayad Kuflik, Tsvi Shimshoni, Ilan Sensors (Basel) Article In this paper, we present a framework for 3D gaze estimation intended to identify the user’s focus of attention in a corneal imaging system. The framework uses a headset that consists of three cameras, a scene camera and two eye cameras: an IR camera and an RGB camera. The IR camera is used to continuously and reliably track the pupil and the RGB camera is used to acquire corneal images of the same eye. Deep learning algorithms are trained to detect the pupil in IR and RGB images and to compute a per user 3D model of the eye in real time. Once the 3D model is built, the 3D gaze direction is computed starting from the eyeball center and passing through the pupil center to the outside world. This model can also be used to transform the pupil position detected in the IR image into its corresponding position in the RGB image and to detect the gaze direction in the corneal image. This technique circumvents the problem of pupil detection in RGB images, which is especially difficult and unreliable when the scene is reflected in the corneal images. In our approach, the auto-calibration process is transparent and unobtrusive. Users do not have to be instructed to look at specific objects to calibrate the eye tracker. They need only to act and gaze normally. The framework was evaluated in a user study in realistic settings and the results are promising. It achieved a very low 3D gaze error (2.12°) and very high accuracy in acquiring corneal images (intersection over union— [Formula: see text] = 0.71). The framework may be used in a variety of real-world mobile scenarios (indoors, indoors near windows and outdoors) with high accuracy. MDPI 2022-12-29 /pmc/articles/PMC9823916/ /pubmed/36616978 http://dx.doi.org/10.3390/s23010381 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Mokatren, Moayad Kuflik, Tsvi Shimshoni, Ilan 3D Gaze Estimation Using RGB-IR Cameras |
title | 3D Gaze Estimation Using RGB-IR Cameras |
title_full | 3D Gaze Estimation Using RGB-IR Cameras |
title_fullStr | 3D Gaze Estimation Using RGB-IR Cameras |
title_full_unstemmed | 3D Gaze Estimation Using RGB-IR Cameras |
title_short | 3D Gaze Estimation Using RGB-IR Cameras |
title_sort | 3d gaze estimation using rgb-ir cameras |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823916/ https://www.ncbi.nlm.nih.gov/pubmed/36616978 http://dx.doi.org/10.3390/s23010381 |
work_keys_str_mv | AT mokatrenmoayad 3dgazeestimationusingrgbircameras AT kufliktsvi 3dgazeestimationusingrgbircameras AT shimshoniilan 3dgazeestimationusingrgbircameras |