Cargando…
Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propo...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8401047/ https://www.ncbi.nlm.nih.gov/pubmed/34450761 http://dx.doi.org/10.3390/s21165322 |
_version_ | 1783745459531022336 |
---|---|
author | Jiang, Jiu Wang, Xianpei Li, Bowen Tian, Meng Yao, Hongtai |
author_facet | Jiang, Jiu Wang, Xianpei Li, Bowen Tian, Meng Yao, Hongtai |
author_sort | Jiang, Jiu |
collection | PubMed |
description | Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propose a no-reference video quality assessment (NR-VQA) method that adds the enhanced awareness of dynamic information to the perception of static objects. Specifically, we use convolutional networks with different dimensions to extract low-level static-dynamic fusion features for video clips and subsequently implement alignment, followed by a temporal memory module consisting of recurrent neural networks branches and fully connected (FC) branches to construct feature associations in a time series. Meanwhile, in order to simulate human visual habits, we built a parametric adaptive network structure to obtain the final score. We further validated the proposed method on four datasets (CVD2014, KoNViD-1k, LIVE-Qualcomm, and LIVE-VQC) to test the generalization ability. Extensive experiments have demonstrated that the proposed method not only outperforms other NR-VQA methods in terms of overall performance of mixed datasets but also achieves competitive performance in individual datasets compared to the existing state-of-the-art methods. |
format | Online Article Text |
id | pubmed-8401047 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-84010472021-08-29 Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos Jiang, Jiu Wang, Xianpei Li, Bowen Tian, Meng Yao, Hongtai Sensors (Basel) Article Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propose a no-reference video quality assessment (NR-VQA) method that adds the enhanced awareness of dynamic information to the perception of static objects. Specifically, we use convolutional networks with different dimensions to extract low-level static-dynamic fusion features for video clips and subsequently implement alignment, followed by a temporal memory module consisting of recurrent neural networks branches and fully connected (FC) branches to construct feature associations in a time series. Meanwhile, in order to simulate human visual habits, we built a parametric adaptive network structure to obtain the final score. We further validated the proposed method on four datasets (CVD2014, KoNViD-1k, LIVE-Qualcomm, and LIVE-VQC) to test the generalization ability. Extensive experiments have demonstrated that the proposed method not only outperforms other NR-VQA methods in terms of overall performance of mixed datasets but also achieves competitive performance in individual datasets compared to the existing state-of-the-art methods. MDPI 2021-08-06 /pmc/articles/PMC8401047/ /pubmed/34450761 http://dx.doi.org/10.3390/s21165322 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Jiang, Jiu Wang, Xianpei Li, Bowen Tian, Meng Yao, Hongtai Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title | Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title_full | Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title_fullStr | Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title_full_unstemmed | Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title_short | Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos |
title_sort | multi-dimensional feature fusion network for no-reference quality assessment of in-the-wild videos |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8401047/ https://www.ncbi.nlm.nih.gov/pubmed/34450761 http://dx.doi.org/10.3390/s21165322 |
work_keys_str_mv | AT jiangjiu multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos AT wangxianpei multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos AT libowen multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos AT tianmeng multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos AT yaohongtai multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos |