Cargando…
MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion
Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700337/ https://www.ncbi.nlm.nih.gov/pubmed/34945998 http://dx.doi.org/10.3390/e23121692 |
_version_ | 1784620732682076160 |
---|---|
author | Zhao, Lei Zhang, Mingcheng Ding, Hongwei Cui, Xiaohui |
author_facet | Zhao, Lei Zhang, Mingcheng Ding, Hongwei Cui, Xiaohui |
author_sort | Zhao, Lei |
collection | PubMed |
description | Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods based on forged features have been proposed. Among the popular forged features, textural features are widely used. However, most of the current texture-based detection methods extract textures directly from RGB images, ignoring the mature spectral analysis methods. Therefore, this research proposes a deepfake detection network fusing RGB features and textural information extracted by neural networks and signal processing methods, namely, MFF-Net. Specifically, it consists of four key components: (1) a feature extraction module to further extract textural and frequency information using the Gabor convolution and residual attention blocks; (2) a texture enhancement module to zoom into the subtle textural features in shallow layers; (3) an attention module to force the classifier to focus on the forged part; (4) two instances of feature fusion to firstly fuse textural features from the shallow RGB branch and feature extraction module and then to fuse the textural features and semantic information. Moreover, we further introduce a new diversity loss to force the feature extraction module to learn features of different scales and directions. The experimental results show that MFF-Net has excellent generalization and has achieved state-of-the-art performance on various deepfake datasets. |
format | Online Article Text |
id | pubmed-8700337 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87003372021-12-24 MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion Zhao, Lei Zhang, Mingcheng Ding, Hongwei Cui, Xiaohui Entropy (Basel) Article Significant progress has been made in generating counterfeit images and videos. Forged videos generated by deepfaking have been widely spread and have caused severe societal impacts, which stir up public concern about automatic deepfake detection technology. Recently, many deepfake detection methods based on forged features have been proposed. Among the popular forged features, textural features are widely used. However, most of the current texture-based detection methods extract textures directly from RGB images, ignoring the mature spectral analysis methods. Therefore, this research proposes a deepfake detection network fusing RGB features and textural information extracted by neural networks and signal processing methods, namely, MFF-Net. Specifically, it consists of four key components: (1) a feature extraction module to further extract textural and frequency information using the Gabor convolution and residual attention blocks; (2) a texture enhancement module to zoom into the subtle textural features in shallow layers; (3) an attention module to force the classifier to focus on the forged part; (4) two instances of feature fusion to firstly fuse textural features from the shallow RGB branch and feature extraction module and then to fuse the textural features and semantic information. Moreover, we further introduce a new diversity loss to force the feature extraction module to learn features of different scales and directions. The experimental results show that MFF-Net has excellent generalization and has achieved state-of-the-art performance on various deepfake datasets. MDPI 2021-12-17 /pmc/articles/PMC8700337/ /pubmed/34945998 http://dx.doi.org/10.3390/e23121692 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhao, Lei Zhang, Mingcheng Ding, Hongwei Cui, Xiaohui MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title | MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title_full | MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title_fullStr | MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title_full_unstemmed | MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title_short | MFF-Net: Deepfake Detection Network Based on Multi-Feature Fusion |
title_sort | mff-net: deepfake detection network based on multi-feature fusion |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700337/ https://www.ncbi.nlm.nih.gov/pubmed/34945998 http://dx.doi.org/10.3390/e23121692 |
work_keys_str_mv | AT zhaolei mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion AT zhangmingcheng mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion AT dinghongwei mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion AT cuixiaohui mffnetdeepfakedetectionnetworkbasedonmultifeaturefusion |