Cargando…

Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos

Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propo...

Descripción completa

Detalles Bibliográficos
Autores principales: Jiang, Jiu, Wang, Xianpei, Li, Bowen, Tian, Meng, Yao, Hongtai
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8401047/
https://www.ncbi.nlm.nih.gov/pubmed/34450761
http://dx.doi.org/10.3390/s21165322
_version_ 1783745459531022336
author Jiang, Jiu
Wang, Xianpei
Li, Bowen
Tian, Meng
Yao, Hongtai
author_facet Jiang, Jiu
Wang, Xianpei
Li, Bowen
Tian, Meng
Yao, Hongtai
author_sort Jiang, Jiu
collection PubMed
description Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propose a no-reference video quality assessment (NR-VQA) method that adds the enhanced awareness of dynamic information to the perception of static objects. Specifically, we use convolutional networks with different dimensions to extract low-level static-dynamic fusion features for video clips and subsequently implement alignment, followed by a temporal memory module consisting of recurrent neural networks branches and fully connected (FC) branches to construct feature associations in a time series. Meanwhile, in order to simulate human visual habits, we built a parametric adaptive network structure to obtain the final score. We further validated the proposed method on four datasets (CVD2014, KoNViD-1k, LIVE-Qualcomm, and LIVE-VQC) to test the generalization ability. Extensive experiments have demonstrated that the proposed method not only outperforms other NR-VQA methods in terms of overall performance of mixed datasets but also achieves competitive performance in individual datasets compared to the existing state-of-the-art methods.
format Online
Article
Text
id pubmed-8401047
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-84010472021-08-29 Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos Jiang, Jiu Wang, Xianpei Li, Bowen Tian, Meng Yao, Hongtai Sensors (Basel) Article Over the past few decades, video quality assessment (VQA) has become a valuable research field. The perception of in-the-wild video quality without reference is mainly challenged by hybrid distortions with dynamic variations and the movement of the content. In order to address this barrier, we propose a no-reference video quality assessment (NR-VQA) method that adds the enhanced awareness of dynamic information to the perception of static objects. Specifically, we use convolutional networks with different dimensions to extract low-level static-dynamic fusion features for video clips and subsequently implement alignment, followed by a temporal memory module consisting of recurrent neural networks branches and fully connected (FC) branches to construct feature associations in a time series. Meanwhile, in order to simulate human visual habits, we built a parametric adaptive network structure to obtain the final score. We further validated the proposed method on four datasets (CVD2014, KoNViD-1k, LIVE-Qualcomm, and LIVE-VQC) to test the generalization ability. Extensive experiments have demonstrated that the proposed method not only outperforms other NR-VQA methods in terms of overall performance of mixed datasets but also achieves competitive performance in individual datasets compared to the existing state-of-the-art methods. MDPI 2021-08-06 /pmc/articles/PMC8401047/ /pubmed/34450761 http://dx.doi.org/10.3390/s21165322 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jiang, Jiu
Wang, Xianpei
Li, Bowen
Tian, Meng
Yao, Hongtai
Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title_full Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title_fullStr Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title_full_unstemmed Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title_short Multi-Dimensional Feature Fusion Network for No-Reference Quality Assessment of In-the-Wild Videos
title_sort multi-dimensional feature fusion network for no-reference quality assessment of in-the-wild videos
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8401047/
https://www.ncbi.nlm.nih.gov/pubmed/34450761
http://dx.doi.org/10.3390/s21165322
work_keys_str_mv AT jiangjiu multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos
AT wangxianpei multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos
AT libowen multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos
AT tianmeng multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos
AT yaohongtai multidimensionalfeaturefusionnetworkfornoreferencequalityassessmentofinthewildvideos