Cargando…
Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module †
Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181782/ https://www.ncbi.nlm.nih.gov/pubmed/37177715 http://dx.doi.org/10.3390/s23094511 |
_version_ | 1785041656794316800 |
---|---|
author | Kuo, Tien-Ying Wei, Yu-Jen Su, Po-Chyi Chao, Chang-Hao |
author_facet | Kuo, Tien-Ying Wei, Yu-Jen Su, Po-Chyi Chao, Chang-Hao |
author_sort | Kuo, Tien-Ying |
collection | PubMed |
description | Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration method to address this issue, which can handle varying degrees of compression artifacts with a single model by predicting the difference between the original and compressed video frames to restore video quality. To achieve this, we adopted a recursive neural network model with dilated convolution, which increases the receptive field of the model while keeping the number of parameters low, making it suitable for deployment on a variety of hardware devices. We also designed a temporal fusion module and integrated the color channels into the objective function. This enables the model to analyze temporal correlation and repair chromaticity artifacts. Despite handling color channels, and unlike other methods that have to train a different model for each quantization parameter (QP), the number of parameters in our lightweight model is kept to only about 269 k, requiring only about one-twelfth of the parameters used by other methods. Our model applied to the HEVC test model (HM) improves the compressed video quality by an average of 0.18 dB of BD-PSNR and −5.06% of BD-BR. |
format | Online Article Text |
id | pubmed-10181782 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-101817822023-05-13 Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † Kuo, Tien-Ying Wei, Yu-Jen Su, Po-Chyi Chao, Chang-Hao Sensors (Basel) Article Video compression algorithms are commonly used to reduce the number of bits required to represent a video with a high compression ratio. However, this can result in the loss of content details and visual artifacts that affect the overall quality of the video. We propose a learning-based restoration method to address this issue, which can handle varying degrees of compression artifacts with a single model by predicting the difference between the original and compressed video frames to restore video quality. To achieve this, we adopted a recursive neural network model with dilated convolution, which increases the receptive field of the model while keeping the number of parameters low, making it suitable for deployment on a variety of hardware devices. We also designed a temporal fusion module and integrated the color channels into the objective function. This enables the model to analyze temporal correlation and repair chromaticity artifacts. Despite handling color channels, and unlike other methods that have to train a different model for each quantization parameter (QP), the number of parameters in our lightweight model is kept to only about 269 k, requiring only about one-twelfth of the parameters used by other methods. Our model applied to the HEVC test model (HM) improves the compressed video quality by an average of 0.18 dB of BD-PSNR and −5.06% of BD-BR. MDPI 2023-05-05 /pmc/articles/PMC10181782/ /pubmed/37177715 http://dx.doi.org/10.3390/s23094511 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Kuo, Tien-Ying Wei, Yu-Jen Su, Po-Chyi Chao, Chang-Hao Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title | Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title_full | Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title_fullStr | Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title_full_unstemmed | Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title_short | Improving Compressed Video Using Single Lightweight Model with Temporal Fusion Module † |
title_sort | improving compressed video using single lightweight model with temporal fusion module † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181782/ https://www.ncbi.nlm.nih.gov/pubmed/37177715 http://dx.doi.org/10.3390/s23094511 |
work_keys_str_mv | AT kuotienying improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule AT weiyujen improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule AT supochyi improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule AT chaochanghao improvingcompressedvideousingsinglelightweightmodelwithtemporalfusionmodule |