Cargando…

Exposing Manipulated Photos and Videos in Digital Forensics Analysis

Tampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and d...

Descripción completa

Detalles Bibliográficos
Autores principales: Ferreira, Sara, Antunes, Mário, Correia, Manuel E.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8321380/
http://dx.doi.org/10.3390/jimaging7070102
_version_ 1783730838842638336
author Ferreira, Sara
Antunes, Mário
Correia, Manuel E.
author_facet Ferreira, Sara
Antunes, Mário
Correia, Manuel E.
author_sort Ferreira, Sara
collection PubMed
description Tampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and disseminating medium. Criminal investigation has been challenged in applying machine learning techniques to automatically distinguish between fake and genuine seized photos and videos. Despite the pertinent need for manual validation, easy-to-use platforms for digital forensics are essential to automate and facilitate the detection of tampered content and to help criminal investigators with their work. This paper presents a machine learning Support Vector Machines (SVM) based method to distinguish between genuine and fake multimedia files, namely digital photos and videos, which may indicate the presence of deepfake content. The method was implemented in Python and integrated as new modules in the widely used digital forensics application Autopsy. The implemented approach extracts a set of simple features resulting from the application of a Discrete Fourier Transform (DFT) to digital photos and video frames. The model was evaluated with a large dataset of classified multimedia files containing both legitimate and fake photos and frames extracted from videos. Regarding deepfake detection in videos, the Celeb-DFv1 dataset was used, featuring 590 original videos collected from YouTube, and covering different subjects. The results obtained with the 5-fold cross-validation outperformed those SVM-based methods documented in the literature, by achieving an average F1-score of [Formula: see text] , [Formula: see text] , and [Formula: see text] , respectively for photos, videos, and a mixture of both types of content. A benchmark with state-of-the-art methods was also done, by comparing the proposed SVM method with deep learning approaches, namely Convolutional Neural Networks (CNN). Despite CNN having outperformed the proposed DFT-SVM compound method, the competitiveness of the results attained by DFT-SVM and the substantially reduced processing time make it appropriate to be implemented and embedded into Autopsy modules, by predicting the level of fakeness calculated for each analyzed multimedia file.
format Online
Article
Text
id pubmed-8321380
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83213802021-08-26 Exposing Manipulated Photos and Videos in Digital Forensics Analysis Ferreira, Sara Antunes, Mário Correia, Manuel E. J Imaging Article Tampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and disseminating medium. Criminal investigation has been challenged in applying machine learning techniques to automatically distinguish between fake and genuine seized photos and videos. Despite the pertinent need for manual validation, easy-to-use platforms for digital forensics are essential to automate and facilitate the detection of tampered content and to help criminal investigators with their work. This paper presents a machine learning Support Vector Machines (SVM) based method to distinguish between genuine and fake multimedia files, namely digital photos and videos, which may indicate the presence of deepfake content. The method was implemented in Python and integrated as new modules in the widely used digital forensics application Autopsy. The implemented approach extracts a set of simple features resulting from the application of a Discrete Fourier Transform (DFT) to digital photos and video frames. The model was evaluated with a large dataset of classified multimedia files containing both legitimate and fake photos and frames extracted from videos. Regarding deepfake detection in videos, the Celeb-DFv1 dataset was used, featuring 590 original videos collected from YouTube, and covering different subjects. The results obtained with the 5-fold cross-validation outperformed those SVM-based methods documented in the literature, by achieving an average F1-score of [Formula: see text] , [Formula: see text] , and [Formula: see text] , respectively for photos, videos, and a mixture of both types of content. A benchmark with state-of-the-art methods was also done, by comparing the proposed SVM method with deep learning approaches, namely Convolutional Neural Networks (CNN). Despite CNN having outperformed the proposed DFT-SVM compound method, the competitiveness of the results attained by DFT-SVM and the substantially reduced processing time make it appropriate to be implemented and embedded into Autopsy modules, by predicting the level of fakeness calculated for each analyzed multimedia file. MDPI 2021-06-24 /pmc/articles/PMC8321380/ http://dx.doi.org/10.3390/jimaging7070102 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ferreira, Sara
Antunes, Mário
Correia, Manuel E.
Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title_full Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title_fullStr Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title_full_unstemmed Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title_short Exposing Manipulated Photos and Videos in Digital Forensics Analysis
title_sort exposing manipulated photos and videos in digital forensics analysis
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8321380/
http://dx.doi.org/10.3390/jimaging7070102
work_keys_str_mv AT ferreirasara exposingmanipulatedphotosandvideosindigitalforensicsanalysis
AT antunesmario exposingmanipulatedphotosandvideosindigitalforensicsanalysis
AT correiamanuele exposingmanipulatedphotosandvideosindigitalforensicsanalysis