Logo do repositório
 
Publicação

Exposing Manipulated Photos and Videos in Digital Forensics Analysis

datacite.subject.fosCiências Médicas::Medicina Clínica
datacite.subject.fosCiências Naturais::Ciências da Computação e da Informação
datacite.subject.fosEngenharia e Tecnologia::Engenharia Eletrotécnica, Eletrónica e Informática
datacite.subject.sdg08:Trabalho Digno e Crescimento Económico
datacite.subject.sdg09:Indústria, Inovação e Infraestruturas
datacite.subject.sdg10:Reduzir as Desigualdades
dc.contributor.authorFerreira, Sara
dc.contributor.authorAntunes, Mário
dc.contributor.authorCorreia, Manuel E.
dc.date.accessioned2026-02-23T16:09:52Z
dc.date.available2026-02-23T16:09:52Z
dc.date.issued2021-06-24
dc.description.abstractTampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and disseminating medium. Criminal investigation has been challenged in applying machine learning techniques to automatically distinguish between fake and genuine seized photos and videos. Despite the pertinent need for manual validation, easy-to-use platforms for digital forensics are essential to automate and facilitate the detection of tampered content and to help criminal investigators with their work. This paper presents a machine learning Support Vector Machines (SVM) based method to distinguish between genuine and fake multimedia files, namely digital photos and videos, which may indicate the presence of deepfake content. The method was implemented in Python and integrated as new modules in the widely used digital forensics application Autopsy. The implemented approach extracts a set of simple features resulting from the application of a Discrete Fourier Transform (DFT) to digital photos and video frames. The model was evaluated with a large dataset of classified multimedia files containing both legitimate and fake photos and frames extracted from videos. Regarding deepfake detection in videos, the Celeb-DFv1 dataset was used, featuring 590 original videos collected from YouTube, and covering different subjects. The results obtained with the 5-fold cross-validation outperformed those SVM-based methods documented in the literature, by achieving an average F1-score of 99.53%, 79.55%, and 89.10%, respectively for photos, videos, and a mixture of both types of content. A benchmark with state-of-the-art methods was also done, by comparing the proposed SVM method with deep learning approaches, namely Convolutional Neural Networks (CNN). Despite CNN having outperformed the proposed DFT-SVM compound method, the competitiveness of the results attained by DFT-SVM and the substantially reduced processing time make it appropriate to be implemented and embedded into Autopsy modules, by predicting the level of fakeness calculated for each analyzed multimedia file.eng
dc.description.sponsorshipThe authors acknowledge the facilities provided by INESC TEC, Faculty of Sciences, and University of Porto, for the support to this research. This work is financed by National Funds through the Portuguese funding agency, FCT - Fundação para a Ciência e a Tecnologia, within project UIDB/50014/2020.
dc.identifier.citationFerreira, S.; Antunes, M.; Correia, M.E. Exposing Manipulated Photos and Videos in Digital Forensics Analysis. J. Imaging 2021, 7, 102. https://doi.org/10.3390/jimaging7070102.
dc.identifier.doi10.3390/jimaging7070102
dc.identifier.eissn2313-433X
dc.identifier.urihttp://hdl.handle.net/10400.8/15699
dc.language.isoeng
dc.peerreviewedyes
dc.publisherMDPI
dc.relationINESC TEC- Institute for Systems and Computer Engineering, Technology and Science
dc.relation.hasversionhttps://www.mdpi.com/2313-433X/7/7/102
dc.relation.ispartofJournal of Imaging
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/
dc.subjectdigital forensics
dc.subjectcybersecurity
dc.subjectmultimedia content manipulation
dc.subjectdeepfake
dc.subjectconvolutional neural networks
dc.subjectsupport vector machines
dc.subjectdiscrete fourier transform
dc.titleExposing Manipulated Photos and Videos in Digital Forensics Analysiseng
dc.typejournal article
dspace.entity.typePublication
oaire.awardTitleINESC TEC- Institute for Systems and Computer Engineering, Technology and Science
oaire.awardURIinfo:eu-repo/grantAgreement/FCT/6817 - DCRRNI ID/UIDB%2F50014%2F2020/PT
oaire.citation.endPage23
oaire.citation.issue7
oaire.citation.startPage1
oaire.citation.titleJournal of Imaging
oaire.citation.volume7
oaire.fundingStream6817 - DCRRNI ID
oaire.versionhttp://purl.org/coar/version/c_970fb48d4fbd8a85
person.familyNameAntunes
person.givenNameMário
person.identifierR-000-NX4
person.identifier.ciencia-idAF10-7EDD-5153
person.identifier.orcid0000-0003-3448-6726
person.identifier.scopus-author-id25930820200
project.funder.identifierhttp://doi.org/10.13039/501100001871
project.funder.nameFundação para a Ciência e a Tecnologia
relation.isAuthorOfPublicatione3e87fb0-d1d6-44c3-985d-920a5560f8c1
relation.isAuthorOfPublication.latestForDiscoverye3e87fb0-d1d6-44c3-985d-920a5560f8c1
relation.isProjectOfPublication42e9aded-d47f-4d1e-a69d-f5566c0b595d
relation.isProjectOfPublication.latestForDiscovery42e9aded-d47f-4d1e-a69d-f5566c0b595d

Ficheiros

Principais
A mostrar 1 - 1 de 1
A carregar...
Miniatura
Nome:
Exposing manipulated photos and videos in digital forensics analysis.pdf
Tamanho:
11.19 MB
Formato:
Adobe Portable Document Format
Descrição:
Tampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and disseminating medium. Criminal investigation has been challenged in applying machine learning techniques to automatically distinguish between fake and genuine seized photos and videos. Despite the pertinent need for manual validation, easy-to-use platforms for digital forensics are essential to automate and facilitate the detection of tampered content and to help criminal investigators with their work. This paper presents a machine learning Support Vector Machines (SVM) based method to distinguish between genuine and fake multimedia files, namely digital photos and videos, which may indicate the presence of deepfake content. The method was implemented in Python and integrated as new modules in the widely used digital forensics application Autopsy. The implemented approach extracts a set of simple features resulting from the application of a Discrete Fourier Transform (DFT) to digital photos and video frames. The model was evaluated with a large dataset of classified multimedia files containing both legitimate and fake photos and frames extracted from videos. Regarding deepfake detection in videos, the Celeb-DFv1 dataset was used, featuring 590 original videos collected from YouTube, and covering different subjects. The results obtained with the 5-fold cross-validation outperformed those SVM-based methods documented in the literature, by achieving an average F1-score of 99.53%, 79.55%, and 89.10%, respectively for photos, videos, and a mixture of both types of content. A benchmark with state-of-the-art methods was also done, by comparing the proposed SVM method with deep learning approaches, namely Convolutional Neural Networks (CNN). Despite CNN having outperformed the proposed DFT-SVM compound method, the competitiveness of the results attained by DFT-SVM and the substantially reduced processing time make it appropriate to be implemented and embedded into Autopsy modules, by predicting the level of fakeness calculated for each analyzed multimedia file.
Licença
A mostrar 1 - 1 de 1
Miniatura indisponível
Nome:
license.txt
Tamanho:
1.32 KB
Formato:
Item-specific license agreed upon to submission
Descrição: