| Literature DB >> 35721408 |
Venkatachalam Kandasamy1, Štěpán Hubálovský1, Pavel Trojovský2.
Abstract
Deepfake (DF) is a kind of forged image or video that is developed to spread misinformation and facilitate vulnerabilities to privacy hacking and truth masking with advanced technologies, including deep learning and artificial intelligence with trained algorithms. This kind of multimedia manipulation, such as changing facial expressions or speech, can be used for a variety of purposes to spread misinformation or exploitation. This kind of multimedia manipulation, such as changing facial expressions or speech, can be used for a variety of purposes to spread misinformation or exploitation. With the recent advancement of generative adversarial networks (GANs) in deep learning models, DF has become an essential part of social media. To detect forged video and images, numerous methods have been developed, and those methods are focused on a particular domain and obsolete in the case of new attacks/threats. Hence, a novel method needs to be developed to tackle new attacks. The method introduced in this article can detect various types of spoofs of images and videos that are computationally generated using deep learning models, such as variants of long short-term memory and convolutional neural networks. The first phase of this proposed work extracts the feature frames from the forged video/image using a sparse autoencoder with a graph long short-term memory (SAE-GLSTM) method at training time. The first phase of this proposed work extracts the feature frames from the forged video/image using a sparse autoencoder with a graph long short-term memory (SAE-GLSTM) method at training time. The proposed DF detection model is tested using the FFHQ database, 100K-Faces, Celeb-DF (V2) and WildDeepfake. The evaluated results show the effectiveness of the proposed method.Entities:
Keywords: Capsule convolution neural network; Deep learning; DeepFake; Generative adversarial networks; Graph LSTM; Long short term memory (LSTM)
Year: 2022 PMID: 35721408 PMCID: PMC9202621 DOI: 10.7717/peerj-cs.953
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
Survey on deepfake detection methods.
| Author | Classifier | Type of input | Dataset |
|---|---|---|---|
|
| CNN concatenated to CFFN | Image | CelebA, DCGAN WGAN WGAN-GP, least squares GAN PGGAN. |
|
| Convolutional bidirectional recurrent LSTM network | Videos | FaceForensics++ and Celeb-DF (5,639 deepfake videos) and the ASVSpoof Access audio dataset. |
|
| CNN | Videos | Four in-the-wild lip-sync deep fakes from Instagram and YouTube (www.instagram.com/bill posters ukand youtu.be/VWMEDacz3L4). |
|
| ResNet50model [102], pretrained on VGGFace2 | Videos | VidTIMIT and two other original datasets obtained from the COHFACE and Deepfake TIMIT datasets. |
|
| Spatiotemporal features with RCN | Videos | FaceForensics++ dataset, including 1,000 videos. |
|
| DCGAN, WGAN-GP and PGGAN. | Images | CelebA-HQ, DCGAN, GAN-GP and PGGAN |
|
| SVM | Videos/Images | UADFV consists of 49 deepfake videos, and 252 deepfake images from DARPA MediFor GAN Image/Video Challenge. |
|
| Capsule networks | Videos/Images | The Idiap Research Institute replayattack, facial reenactment FaceForensics. |
|
| CNN | Videos | Deepfake one constituted from onlinevideos and the FaceForensics one created by the Face2Face approach. |
|
| CNN and LSTM | Videos | A collection of 600 videos obtained from multiple websites. |
| LRCN | Videos | Consists of 49 interview and presentation videos, and their corresponding generated deepfakes. |
Figure 1Overview of proposed DeepFake detection system.
Figure 2Proposed stages of preprocessing.
Two-Level Adaptive Median Noise Removal Filter.
| Step 1: Assume that |
| Step 2: Consider a |
| Step 3: Calculate |
| Step 4: Level |
|
|
| Step 5: |
| Step 6: |
| Step 7: |
| Step 8: |
| Step 9: |
| Step 10: |
| Step 11: Level |
|
|
| Step 12: |
| Step 13: |
| Step 14: |
Figure 3Deep fake image feature extraction using proposed SAE-GLSTM network.
Figure 4GLSTM cell.
Figure 5Hierarchical timing structure of (A) forward GLSTM (B) backward GLSTM.
Forward and backward sequence equations of the graph LSTM.
| Forward sequence | Backward sequence |
|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Figure 6Karate club network for LC convolution (Kipf & Welling, 2017).
Figure 7Capsule dual graph CNN structure.
Capsule Dual Graph CNN (C-DGCNN).
| Step 1: |
| Step 2: |
| Step 3: |
| Step 4: Compute Loss using |
| Step 5: |
| Step 6: |
| Step 7: |
| Step 8: Dynamic routing procedure ( |
| Step 9: |
| Step 10: |
| Step 11: |
| Step 12: |
| Step 13: |
| Step 14: |
| Step 15: |
| Step 16: |
| Step 17: |
| Step 18: |
| Step 19: |
| Step 20: |
| Step 21: |
| Step 22: |
| Step 23: |
| Step 24: |
Accuracy comparison of the proposed vs traditional baseline systems for various datasets.
| Methods | Datasets | |||
|---|---|---|---|---|
| FFHQ | 100K-Faces | Celeb-DF | WildDeepfake | |
| VGG19 | 84.5 | 74.12 | 88.43 | 89.25 |
| ResNet | 88.32 | 80.11 | 89.32 | 86.52 |
| MobileNet | 91.15 | 90.21 | 90.01 | 96.75 |
| Proposed SAE-GLSTM-CDGCNN | 96.92 | 97.15 | 98.12 | 98.91 |
Sensitivity and specificity analysis of the proposed system on different datasets.
| Datasets | Sensitivity % | Specificity % | ||||||
|---|---|---|---|---|---|---|---|---|
| VGG19 | ResNet | Mobile- Net | Proposed SAE-GLSTM-CDGCNN | VGG19 | ResNet | Mobile- Net | Proposed SAE-GLSTM-CDGCNN | |
| FFHQ | 87.3 | 81.6 | 84.2 | 91.67 | 84.23 | 86.1 | 85.2 | 92.4 |
| 100K-Faces | 86.2 | 83 | 85.6 | 89.8 | 84.2 | 86.2 | 89.54 | 93.5 |
| Celeb-DF | 84.9 | 88.3 | 83.7 | 89.1 | 87.1 | 87.1 | 89.1 | 94.2 |
| WildDeepfake | 91.2 | 85.3 | 90.1 | 93.1 | 86.1 | 89.4 | 86.3 | 95.2 |
Figure 8ROC value comparison.
Performance comparison of the proposed methods with different datasets in terms of the error detection rate.
| DF detection methods | Datasets | |||
|---|---|---|---|---|
| FFHQ | 100K-Faces | Celeb-DF | WildDeepfake | |
| VGG19 | 13.11 | 18.32 | 12.3 | 11.4 |
| ResNet | 13.4 | 15.2 | 12.1 | 11.2 |
| MobileNet | 11.2 | 14.22 | 11.65 | 9.21 |
| Proposed SAE-GLSTM-CDGCNN | 5.91 | 6.01 | 7.12 | 5.1 |
Figure 9EER comparison of DF detection systems.
Figure 10Computation time.