Cargando…

Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †

Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration...

Descripción completa

Detalles Bibliográficos
Autores principales: Kuo, Tien-Ying, Wei, Yu-Jen, Su, Po-Chyi, Chao, Chang-Hao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181782/
https://www.ncbi.nlm.nih.gov/pubmed/37177715
http://dx.doi.org/10.3390/s23094511
_version_ 1785041656794316800
author Kuo, Tien-Ying
Wei, Yu-Jen
Su, Po-Chyi
Chao, Chang-Hao
author_facet Kuo, Tien-Ying
Wei, Yu-Jen
Su, Po-Chyi
Chao, Chang-Hao
author_sort Kuo, Tien-Ying
collection PubMed
description Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration method to address this issue, which can handle varying degrees of compression artifacts with a single model by predicting the difference between the original and compressed video frames to restore video quality. To achieve this, we adopted a recursive neural network model with dilated convolution, which increases the receptive field of the model while keeping the number of parameters low, making it suitable for deployment on a variety of hardware devices. We also designed a temporal fusion module and integrated the color channels into the objective function. This enables the model to analyze temporal correlation and repair chromaticity artifacts. Despite handling color channels, and unlike other methods that have to train a different model for each quantization parameter (QP), the number of parameters in our lightweight model is kept to only about 269 k, requiring only about one-twelfth of the parameters used by other methods. Our model applied to the HEVC test model (HM) improves the compressed video quality by an average of 0.18 dB of BD-PSNR and −5.06% of BD-BR.
format Online
Article
Text
id pubmed-10181782
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-101817822023-05-13 Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † Kuo, Tien-Ying Wei, Yu-Jen Su, Po-Chyi Chao, Chang-Hao Sensors (Basel) Article Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration method to address this issue, which can handle varying degrees of compression artifacts with a single model by predicting the difference between the original and compressed video frames to restore video quality. To achieve this, we adopted a recursive neural network model with dilated convolution, which increases the receptive field of the model while keeping the number of parameters low, making it suitable for deployment on a variety of hardware devices. We also designed a temporal fusion module and integrated the color channels into the objective function. This enables the model to analyze temporal correlation and repair chromaticity artifacts. Despite handling color channels, and unlike other methods that have to train a different model for each quantization parameter (QP), the number of parameters in our lightweight model is kept to only about 269 k, requiring only about one-twelfth of the parameters used by other methods. Our model applied to the HEVC test model (HM) improves the compressed video quality by an average of 0.18 dB of BD-PSNR and −5.06% of BD-BR. MDPI 2023-05-05 /pmc/articles/PMC10181782/ /pubmed/37177715 http://dx.doi.org/10.3390/s23094511 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kuo, Tien-Ying
Wei, Yu-Jen
Su, Po-Chyi
Chao, Chang-Hao
Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title_full Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title_fullStr Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title_full_unstemmed Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title_short Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
title_sort improving compressed video using single lightweight model with temporal fusion module †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181782/
https://www.ncbi.nlm.nih.gov/pubmed/37177715
http://dx.doi.org/10.3390/s23094511
work_keys_str_mv AT kuotienying improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule
AT weiyujen improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule
AT supochyi improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule
AT chaochanghao improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule