Cargando…
Deepfake video detection: YOLO-Face convolution recurrent approach
Recently, the deepfake techniques for swapping faces have been spreading, allowing easy creation of hyper-realistic fake videos. Detecting the authenticity of a video has become increasingly critical because of the potential negative impact on the world. Here, a new project is introduced; You Only L...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8507472/ https://www.ncbi.nlm.nih.gov/pubmed/34712799 http://dx.doi.org/10.7717/peerj-cs.730 |
_version_ | 1784581863725072384 |
---|---|
author | Ismail, Aya Elpeltagy, Marwa Zaki, Mervat ElDahshan, Kamal A. |
author_facet | Ismail, Aya Elpeltagy, Marwa Zaki, Mervat ElDahshan, Kamal A. |
author_sort | Ismail, Aya |
collection | PubMed |
description | Recently, the deepfake techniques for swapping faces have been spreading, allowing easy creation of hyper-realistic fake videos. Detecting the authenticity of a video has become increasingly critical because of the potential negative impact on the world. Here, a new project is introduced; You Only Look Once Convolution Recurrent Neural Networks (YOLO-CRNNs), to detect deepfake videos. The YOLO-Face detector detects face regions from each frame in the video, whereas a fine-tuned EfficientNet-B5 is used to extract the spatial features of these faces. These features are fed as a batch of input sequences into a Bidirectional Long Short-Term Memory (Bi-LSTM), to extract the temporal features. The new scheme is then evaluated on a new large-scale dataset; CelebDF-FaceForencics++ (c23), based on a combination of two popular datasets; FaceForencies++ (c23) and Celeb-DF. It achieves an Area Under the Receiver Operating Characteristic Curve (AUROC) 89.35% score, 89.38% accuracy, 83.15% recall, 85.55% precision, and 84.33% F1-measure for pasting data approach. The experimental analysis approves the superiority of the proposed method compared to the state-of-the-art methods. |
format | Online Article Text |
id | pubmed-8507472 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-85074722021-10-27 Deepfake video detection: YOLO-Face convolution recurrent approach Ismail, Aya Elpeltagy, Marwa Zaki, Mervat ElDahshan, Kamal A. PeerJ Comput Sci Artificial Intelligence Recently, the deepfake techniques for swapping faces have been spreading, allowing easy creation of hyper-realistic fake videos. Detecting the authenticity of a video has become increasingly critical because of the potential negative impact on the world. Here, a new project is introduced; You Only Look Once Convolution Recurrent Neural Networks (YOLO-CRNNs), to detect deepfake videos. The YOLO-Face detector detects face regions from each frame in the video, whereas a fine-tuned EfficientNet-B5 is used to extract the spatial features of these faces. These features are fed as a batch of input sequences into a Bidirectional Long Short-Term Memory (Bi-LSTM), to extract the temporal features. The new scheme is then evaluated on a new large-scale dataset; CelebDF-FaceForencics++ (c23), based on a combination of two popular datasets; FaceForencies++ (c23) and Celeb-DF. It achieves an Area Under the Receiver Operating Characteristic Curve (AUROC) 89.35% score, 89.38% accuracy, 83.15% recall, 85.55% precision, and 84.33% F1-measure for pasting data approach. The experimental analysis approves the superiority of the proposed method compared to the state-of-the-art methods. PeerJ Inc. 2021-09-21 /pmc/articles/PMC8507472/ /pubmed/34712799 http://dx.doi.org/10.7717/peerj-cs.730 Text en © 2021 Ismail et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Artificial Intelligence Ismail, Aya Elpeltagy, Marwa Zaki, Mervat ElDahshan, Kamal A. Deepfake video detection: YOLO-Face convolution recurrent approach |
title | Deepfake video detection: YOLO-Face convolution recurrent approach |
title_full | Deepfake video detection: YOLO-Face convolution recurrent approach |
title_fullStr | Deepfake video detection: YOLO-Face convolution recurrent approach |
title_full_unstemmed | Deepfake video detection: YOLO-Face convolution recurrent approach |
title_short | Deepfake video detection: YOLO-Face convolution recurrent approach |
title_sort | deepfake video detection: yolo-face convolution recurrent approach |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8507472/ https://www.ncbi.nlm.nih.gov/pubmed/34712799 http://dx.doi.org/10.7717/peerj-cs.730 |
work_keys_str_mv | AT ismailaya deepfakevideodetectionyolofaceconvolutionrecurrentapproach AT elpeltagymarwa deepfakevideodetectionyolofaceconvolutionrecurrentapproach AT zakimervat deepfakevideodetectionyolofaceconvolutionrecurrentapproach AT eldahshankamala deepfakevideodetectionyolofaceconvolutionrecurrentapproach |