Cargando…
Deep Sensing for Compressive Video Acquisition †
A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but b...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10490772/ https://www.ncbi.nlm.nih.gov/pubmed/37687990 http://dx.doi.org/10.3390/s23177535 |
_version_ | 1785103918262386688 |
---|---|
author | Yoshida, Michitaka Torii, Akihiko Okutomi, Masatoshi Taniguchi, Rin-ichiro Nagahara, Hajime Yagi, Yasushi |
author_facet | Yoshida, Michitaka Torii, Akihiko Okutomi, Masatoshi Taniguchi, Rin-ichiro Nagahara, Hajime Yagi, Yasushi |
author_sort | Yoshida, Michitaka |
collection | PubMed |
description | A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but by optimizing the sensing matrix, we can capture images more efficiently and reconstruct multidimensional information with high quality. Although compressive video sensing requires random sampling as a theoretical optimum, when designing the sensing matrix in practice, there are many hardware limitations (such as exposure and color filter patterns). Existing studies have found random sampling is not always the best solution for compressive sensing because the optimal sampling pattern is related to the scene context, and it is hard to manually design a sampling pattern and reconstruction algorithm. In this paper, we propose an end-to-end learning approach that jointly optimizes the sampling pattern as well as the reconstruction decoder. We applied this deep sensing approach to the video compressive sensing problem. We modeled the spatio–temporal sampling and color filter pattern using a convolutional neural network constrained by hardware limitations during network training. We demonstrated that the proposed method performs better than the manually designed method in gray-scale video and color video acquisitions. |
format | Online Article Text |
id | pubmed-10490772 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-104907722023-09-09 Deep Sensing for Compressive Video Acquisition † Yoshida, Michitaka Torii, Akihiko Okutomi, Masatoshi Taniguchi, Rin-ichiro Nagahara, Hajime Yagi, Yasushi Sensors (Basel) Article A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but by optimizing the sensing matrix, we can capture images more efficiently and reconstruct multidimensional information with high quality. Although compressive video sensing requires random sampling as a theoretical optimum, when designing the sensing matrix in practice, there are many hardware limitations (such as exposure and color filter patterns). Existing studies have found random sampling is not always the best solution for compressive sensing because the optimal sampling pattern is related to the scene context, and it is hard to manually design a sampling pattern and reconstruction algorithm. In this paper, we propose an end-to-end learning approach that jointly optimizes the sampling pattern as well as the reconstruction decoder. We applied this deep sensing approach to the video compressive sensing problem. We modeled the spatio–temporal sampling and color filter pattern using a convolutional neural network constrained by hardware limitations during network training. We demonstrated that the proposed method performs better than the manually designed method in gray-scale video and color video acquisitions. MDPI 2023-08-30 /pmc/articles/PMC10490772/ /pubmed/37687990 http://dx.doi.org/10.3390/s23177535 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yoshida, Michitaka Torii, Akihiko Okutomi, Masatoshi Taniguchi, Rin-ichiro Nagahara, Hajime Yagi, Yasushi Deep Sensing for Compressive Video Acquisition † |
title | Deep Sensing for Compressive Video Acquisition † |
title_full | Deep Sensing for Compressive Video Acquisition † |
title_fullStr | Deep Sensing for Compressive Video Acquisition † |
title_full_unstemmed | Deep Sensing for Compressive Video Acquisition † |
title_short | Deep Sensing for Compressive Video Acquisition † |
title_sort | deep sensing for compressive video acquisition † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10490772/ https://www.ncbi.nlm.nih.gov/pubmed/37687990 http://dx.doi.org/10.3390/s23177535 |
work_keys_str_mv | AT yoshidamichitaka deepsensingforcompressivevideoacquisition AT toriiakihiko deepsensingforcompressivevideoacquisition AT okutomimasatoshi deepsensingforcompressivevideoacquisition AT taniguchirinichiro deepsensingforcompressivevideoacquisition AT nagaharahajime deepsensingforcompressivevideoacquisition AT yagiyasushi deepsensingforcompressivevideoacquisition |