Cargando…
Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras
This paper describes three coarse image description strategies, which are meant to promote a rough perception of surrounding objects for visually impaired individuals, with application to indoor spaces. The described algorithms operate on images (grabbed by the user, by means of a chest-mounted came...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5712811/ https://www.ncbi.nlm.nih.gov/pubmed/29144395 http://dx.doi.org/10.3390/s17112641 |
_version_ | 1783283293099130880 |
---|---|
author | Malek, Salim Melgani, Farid Mekhalfi, Mohamed Lamine Bazi, Yakoub |
author_facet | Malek, Salim Melgani, Farid Mekhalfi, Mohamed Lamine Bazi, Yakoub |
author_sort | Malek, Salim |
collection | PubMed |
description | This paper describes three coarse image description strategies, which are meant to promote a rough perception of surrounding objects for visually impaired individuals, with application to indoor spaces. The described algorithms operate on images (grabbed by the user, by means of a chest-mounted camera), and provide in output a list of objects that likely exist in his context across the indoor scene. In this regard, first, different colour, texture, and shape-based feature extractors are generated, followed by a feature learning step by means of AutoEncoder (AE) models. Second, the produced features are fused and fed into a multilabel classifier in order to list the potential objects. The conducted experiments point out that fusing a set of AE-learned features scores higher classification rates with respect to using the features individually. Furthermore, with respect to reference works, our method: (i) yields higher classification accuracies, and (ii) runs (at least four times) faster, which enables a potential full real-time application. |
format | Online Article Text |
id | pubmed-5712811 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-57128112017-12-07 Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras Malek, Salim Melgani, Farid Mekhalfi, Mohamed Lamine Bazi, Yakoub Sensors (Basel) Article This paper describes three coarse image description strategies, which are meant to promote a rough perception of surrounding objects for visually impaired individuals, with application to indoor spaces. The described algorithms operate on images (grabbed by the user, by means of a chest-mounted camera), and provide in output a list of objects that likely exist in his context across the indoor scene. In this regard, first, different colour, texture, and shape-based feature extractors are generated, followed by a feature learning step by means of AutoEncoder (AE) models. Second, the produced features are fused and fed into a multilabel classifier in order to list the potential objects. The conducted experiments point out that fusing a set of AE-learned features scores higher classification rates with respect to using the features individually. Furthermore, with respect to reference works, our method: (i) yields higher classification accuracies, and (ii) runs (at least four times) faster, which enables a potential full real-time application. MDPI 2017-11-16 /pmc/articles/PMC5712811/ /pubmed/29144395 http://dx.doi.org/10.3390/s17112641 Text en © 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Malek, Salim Melgani, Farid Mekhalfi, Mohamed Lamine Bazi, Yakoub Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title | Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title_full | Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title_fullStr | Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title_full_unstemmed | Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title_short | Real-Time Indoor Scene Description for the Visually Impaired Using Autoencoder Fusion Strategies with Visible Cameras |
title_sort | real-time indoor scene description for the visually impaired using autoencoder fusion strategies with visible cameras |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5712811/ https://www.ncbi.nlm.nih.gov/pubmed/29144395 http://dx.doi.org/10.3390/s17112641 |
work_keys_str_mv | AT maleksalim realtimeindoorscenedescriptionforthevisuallyimpairedusingautoencoderfusionstrategieswithvisiblecameras AT melganifarid realtimeindoorscenedescriptionforthevisuallyimpairedusingautoencoderfusionstrategieswithvisiblecameras AT mekhalfimohamedlamine realtimeindoorscenedescriptionforthevisuallyimpairedusingautoencoderfusionstrategieswithvisiblecameras AT baziyakoub realtimeindoorscenedescriptionforthevisuallyimpairedusingautoencoderfusionstrategieswithvisiblecameras |