Cargando…

MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion

Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Lei, Zhang, Mingcheng, Ding, Hongwei, Cui, Xiaohui
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700337/
https://www.ncbi.nlm.nih.gov/pubmed/34945998
http://dx.doi.org/10.3390/e23121692
_version_ 1784620732682076160
author Zhao, Lei
Zhang, Mingcheng
Ding, Hongwei
Cui, Xiaohui
author_facet Zhao, Lei
Zhang, Mingcheng
Ding, Hongwei
Cui, Xiaohui
author_sort Zhao, Lei
collection PubMed
description Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods based on forged features have been proposed. Among the popular forged features, textural features are widely used. However, most of the current texture-based detection methods extract textures directly from RGB images, ignoring the mature spectral analysis methods. Therefore, this research proposes a deepfake detection network fusing RGB features and textural information extracted by neural networks and signal processing methods, namely, MFF-Net. Specifically, it consists of four key components: (1) a feature extraction module to further extract textural and frequency information using the Gabor convolution and residual attention blocks; (2) a texture enhancement module to zoom into the subtle textural features in shallow layers; (3) an attention module to force the classifier to focus on the forged part; (4) two instances of feature fusion to firstly fuse textural features from the shallow RGB branch and feature extraction module and then to fuse the textural features and semantic information. Moreover, we further introduce a new diversity loss to force the feature extraction module to learn features of different scales and directions. The experimental results show that MFF-Net has excellent generalization and has achieved state-of-the-art performance on various deepfake datasets.
format Online
Article
Text
id pubmed-8700337
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87003372021-12-24 MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion Zhao, Lei Zhang, Mingcheng Ding, Hongwei Cui, Xiaohui Entropy (Basel) Article Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods based on forged features have been proposed. Among the popular forged features, textural features are widely used. However, most of the current texture-based detection methods extract textures directly from RGB images, ignoring the mature spectral analysis methods. Therefore, this research proposes a deepfake detection network fusing RGB features and textural information extracted by neural networks and signal processing methods, namely, MFF-Net. Specifically, it consists of four key components: (1) a feature extraction module to further extract textural and frequency information using the Gabor convolution and residual attention blocks; (2) a texture enhancement module to zoom into the subtle textural features in shallow layers; (3) an attention module to force the classifier to focus on the forged part; (4) two instances of feature fusion to firstly fuse textural features from the shallow RGB branch and feature extraction module and then to fuse the textural features and semantic information. Moreover, we further introduce a new diversity loss to force the feature extraction module to learn features of different scales and directions. The experimental results show that MFF-Net has excellent generalization and has achieved state-of-the-art performance on various deepfake datasets. MDPI 2021-12-17 /pmc/articles/PMC8700337/ /pubmed/34945998 http://dx.doi.org/10.3390/e23121692 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhao, Lei
Zhang, Mingcheng
Ding, Hongwei
Cui, Xiaohui
MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title_full MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title_fullStr MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title_full_unstemmed MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title_short MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
title_sort mff-net: deepfake detection network based on multi-feature fusion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700337/
https://www.ncbi.nlm.nih.gov/pubmed/34945998
http://dx.doi.org/10.3390/e23121692
work_keys_str_mv AT zhaolei mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion
AT zhangmingcheng mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion
AT dinghongwei mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion
AT cuixiaohui mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion