| Literature DB >> 29690629 |
Ye Yao1,2, Weitong Hu3, Wei Zhang4, Ting Wu5, Yun-Qing Shi6.
Abstract
Computer-generated graphics (CGs) are images generated by computer software. The rapid development of computer graphics technologies has made it easier to generate photorealistic computer graphics, and these graphics are quite difficult to distinguish from natural images (NIs) with the naked eye. In this paper, we propose a method based on sensor pattern noise (SPN) and deep learning to distinguish CGs from NIs. Before being fed into our convolutional neural network (CNN)-based model, these images—CGs and NIs—are clipped into image patches. Furthermore, three high-pass filters (HPFs) are used to remove low-frequency signals, which represent the image content. These filters are also used to reveal the residual signal as well as SPN introduced by the digital camera device. Different from the traditional methods of distinguishing CGs from NIs, the proposed method utilizes a five-layer CNN to classify the input image patches. Based on the classification results of the image patches, we deploy a majority vote scheme to obtain the classification results for the full-size images. The experiments have demonstrated that (1) the proposed method with three HPFs can achieve better results than that with only one HPF or no HPF and that (2) the proposed method with three HPFs achieves 100% accuracy, although the NIs undergo a JPEG compression with a quality factor of 75.Entities:
Keywords: computer-generated graphics; convolutional neural network; image forensics; natural images; sensor pattern noise
Year: 2018 PMID: 29690629 PMCID: PMC5948567 DOI: 10.3390/s18041296
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Some examples of computer graphics (CGs) from a dataset in [3].
Figure 2Three high-pass filter (HPFs) used in the proposed method. (a) the high-pass filter of SQUARE 5 × 5, (b) the high-pass filter of EDGE 3 × 3, (c) the high-pass filter of SQUARE 3 × 3.
Figure 3The proposed convolutional neural network architecture. Names and parameters of each layer are displayed in the boxes. Kernel sizes in each convolution layer are shown as number_of_kernels × (width × height × number_of_input). Sizes of feature maps between consecutive layers are shown as number_of_feature_maps × (width × height). Padding is used in each convolutional layer to keep the shape of image patches. BN: batch normalization; ReLUs: rectified linear units.
Figure 4Validation performance of the proposed method.
Figure 5Training loss of the proposed method.
Classification accuracy with different numbers of HPFs.
| Image Patches | Full-Size Images | |||
|---|---|---|---|---|
| Model of 50 Epochs | Model of 80 Epochs | Model of 50 Epochs | Model of 80 Epochs | |
| the proposed HPF × 3 | 99.98% | 99.95% | 100% | 100% |
| the proposed HPF × 1 | 99.87% | 99.77% | 100% | 99.83% |
| the proposed HPF × 0 | 88.28% | 87.77% | 93.37% | 93.12% |
| Rahmouni et al. [ | 84.8% 1 94.4% 2 | 93.2% | ||
1 The size of the image patches was 100 × 100, which is different from the proposed methods. 2 The size of the image patches was 650 × 650. A weighted voting scheme was used.
Figure 6Validation performance of the proposed method.
Figure 7Training loss of the proposed method.
Classification accuracy for different quality factors of natural images (NIs).
| Image Patches | Full-Size Images | |||
|---|---|---|---|---|
| Model of 50 Epochs | Model of 80 Epochs | Model of 50 Epochs | Model of 80 Epochs | |
| 99.52% | 99.71% | 100% | 100% | |
| 99.95% | 99.98% | 100% | 100% | |
| 99.99% | 99.99% | 100% | 100% | |