Cargando…

Deep Sensing for Compressive Video Acquisition †

A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but b...

Descripción completa

Detalles Bibliográficos
Autores principales: Yoshida, Michitaka, Torii, Akihiko, Okutomi, Masatoshi, Taniguchi, Rin-ichiro, Nagahara, Hajime, Yagi, Yasushi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10490772/
https://www.ncbi.nlm.nih.gov/pubmed/37687990
http://dx.doi.org/10.3390/s23177535
_version_ 1785103918262386688
author Yoshida, Michitaka
Torii, Akihiko
Okutomi, Masatoshi
Taniguchi, Rin-ichiro
Nagahara, Hajime
Yagi, Yasushi
author_facet Yoshida, Michitaka
Torii, Akihiko
Okutomi, Masatoshi
Taniguchi, Rin-ichiro
Nagahara, Hajime
Yagi, Yasushi
author_sort Yoshida, Michitaka
collection PubMed
description A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but by optimizing the sensing matrix, we can capture images more efficiently and reconstruct multidimensional information with high quality. Although compressive video sensing requires random sampling as a theoretical optimum, when designing the sensing matrix in practice, there are many hardware limitations (such as exposure and color filter patterns). Existing studies have found random sampling is not always the best solution for compressive sensing because the optimal sampling pattern is related to the scene context, and it is hard to manually design a sampling pattern and reconstruction algorithm. In this paper, we propose an end-to-end learning approach that jointly optimizes the sampling pattern as well as the reconstruction decoder. We applied this deep sensing approach to the video compressive sensing problem. We modeled the spatio–temporal sampling and color filter pattern using a convolutional neural network constrained by hardware limitations during network training. We demonstrated that the proposed method performs better than the manually designed method in gray-scale video and color video acquisitions.
format Online
Article
Text
id pubmed-10490772
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-104907722023-09-09 Deep Sensing for Compressive Video Acquisition † Yoshida, Michitaka Torii, Akihiko Okutomi, Masatoshi Taniguchi, Rin-ichiro Nagahara, Hajime Yagi, Yasushi Sensors (Basel) Article A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but by optimizing the sensing matrix, we can capture images more efficiently and reconstruct multidimensional information with high quality. Although compressive video sensing requires random sampling as a theoretical optimum, when designing the sensing matrix in practice, there are many hardware limitations (such as exposure and color filter patterns). Existing studies have found random sampling is not always the best solution for compressive sensing because the optimal sampling pattern is related to the scene context, and it is hard to manually design a sampling pattern and reconstruction algorithm. In this paper, we propose an end-to-end learning approach that jointly optimizes the sampling pattern as well as the reconstruction decoder. We applied this deep sensing approach to the video compressive sensing problem. We modeled the spatio–temporal sampling and color filter pattern using a convolutional neural network constrained by hardware limitations during network training. We demonstrated that the proposed method performs better than the manually designed method in gray-scale video and color video acquisitions. MDPI 2023-08-30 /pmc/articles/PMC10490772/ /pubmed/37687990 http://dx.doi.org/10.3390/s23177535 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Yoshida, Michitaka
Torii, Akihiko
Okutomi, Masatoshi
Taniguchi, Rin-ichiro
Nagahara, Hajime
Yagi, Yasushi
Deep Sensing for Compressive Video Acquisition †
title Deep Sensing for Compressive Video Acquisition †
title_full Deep Sensing for Compressive Video Acquisition †
title_fullStr Deep Sensing for Compressive Video Acquisition †
title_full_unstemmed Deep Sensing for Compressive Video Acquisition †
title_short Deep Sensing for Compressive Video Acquisition †
title_sort deep sensing for compressive video acquisition †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10490772/
https://www.ncbi.nlm.nih.gov/pubmed/37687990
http://dx.doi.org/10.3390/s23177535
work_keys_str_mv AT yoshidamichitaka deepsensingforcompressivevideoacquisition
AT toriiakihiko deepsensingforcompressivevideoacquisition
AT okutomimasatoshi deepsensingforcompressivevideoacquisition
AT taniguchirinichiro deepsensingforcompressivevideoacquisition
AT nagaharahajime deepsensingforcompressivevideoacquisition
AT yagiyasushi deepsensingforcompressivevideoacquisition