| Literature DB >> 34969837 |
Matthew Groh1, Ziv Epstein2, Chaz Firestone3, Rosalind Picard2.
Abstract
The recent emergence of machine-manipulated media raises an important societal question: How can we know whether a video that we watch is real or fake? In two online studies with 15,016 participants, we present authentic videos and deepfakes and ask participants to identify which is which. We compare the performance of ordinary human observers with the leading computer vision deepfake detection model and find them similarly accurate, while making different kinds of mistakes. Together, participants with access to the model's prediction are more accurate than either alone, but inaccurate model predictions often decrease participants' accuracy. To probe the relative strengths and weaknesses of humans and machines as detectors of deepfakes, we examine human and machine performance across video-level features, and we evaluate the impact of preregistered randomized interventions on deepfake detection. We find that manipulations designed to disrupt visual processing of faces hinder human participants' performance while mostly not affecting the model's performance, suggesting a role for specialized cognitive capacities in explaining human deepfake detection performance.Entities:
Keywords: artificial intelligence; facial recognition; forensic science; misinformation; wisdom of crowds
Mesh:
Year: 2022 PMID: 34969837 PMCID: PMC8740705 DOI: 10.1073/pnas.2110013119
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 11.205
Fig. 1.One of these two images is the first frame of a deepfake from experiment 1; the other is the first frame of the original, authentic video from which the deepfake was created. Experiment 1 asked whether participants can tell which is which, using a two-alternative forced-choice paradigm (i.e., selecting which of two video clips is a deepfake). Experiment 2 presented a single video and asked participants for their confidence the video is a deepfake or not. (Left) The deepfake; the man was not mustachioed at the time of filming. (Right) Authentic.
Fig. 2.(A) The distribution of participant performance across experiments compared to the model’s performance via violin plots where the white dots indicate the mean and the black bars indicate the interquartile range. R, recruited participants; NR, nonrecruited participants; E1, experiment 1; E2, experiment 2. In experiment 1 (two-alternative forced choice), accuracy is defined as identifying a deepfake from a pair of videos correctly. In experiment 2 (single-video design), accurate identification is defined as responding with the correct answer with more than 50% confidence. The model’s performance represents a single observation in each instance, and, as such, we present the model’s performance as a horizontal black line with a white dot in the middle. The crowd mean distributions are obtained by bootstrapping CIs based on 1,000 randomly drawn samples that are each half of the total observations. (B) A scatter plot of the model’s accuracy and the distribution of participants’ accuracy scores for each video. The x axis is an index of the videos, and it is ordered by experiment, true class of each video, and participant’s average accuracy. The teal lines represent the interquartile range of recruited participants’ responses. (C) The distribution of changes in recruited participants’ accuracy after updating their response based on whether the model’s prediction is correct, incorrect, or indecisive. (D) The receiver operator characteristic curves of computer performance, recruited participants’ collective performance, and recruited participants’ collective performance with the model’s decision support across the 50 DFDC videos in experiment 2.
Treatment effects of interventions on accuracy
| Dependent variable: Accuracy | |||||||||
| Recruited | Nonrecruited | Computer | |||||||
| All | Real | Fake | All | Real | Fake | All | Real | Fake | |
| Constant | 0.655*** | 0.716*** | 0.567*** | 0.679*** | 0.700*** | 0.632*** | 0.813*** | 0.786*** | 0.841*** |
| (0.009) | (0.014) | (0.015) | (0.002) | (0.003) | (0.003) | (0.030) | (0.040) | (0.044) | |
| Inversion | –0.043*** | –0.091*** | 0.010 | –0.053*** | –0.080*** | –0.027*** | –0.121*** | –0.110* | –0.132** |
| (0.014) | (0.021) | (0.021) | (0.004) | (0.006) | (0.006) | (0.042) | (0.056) | (0.063) | |
| Misalignment | –0.061*** | –0.042* | –0.077*** | –0.070*** | –0.056*** | –0.084*** | 0.011 | 0.000 | 0.021 |
| (0.016) | (0.024) | (0.025) | (0.005) | (0.007) | (0.007) | (0.042) | (0.056) | (0.063) | |
| Eye occlusion | –0.044*** | –0.023 | –0.063*** | –0.040*** | –0.035*** | –0.043*** | –0.003 | –0.007 | 0.001 |
| (0.015) | (0.021) | (0.024) | (0.004) | (0.006) | (0.006) | (0.042) | (0.056) | (0.063) | |
| Anger | –0.020 | –0.052** | 0.012 | ||||||
| (0.014) | (0.024) | (0.021) | |||||||
| Number of participants | 229 | 229 | 229 | 7,563 | 6,368 | 6,670 | 0 | 0 | 0 |
| Number of guesses (real) | 2,349 | 1,514 | 835 | 27,446 | 18,524 | 8,922 | 81 | 76 | 5 |
| Number of guesses (deepfake) | 1,707 | 549 | 1,158 | 22,766 | 6,316 | 16,450 | 87 | 7 | 80 |
| Number of guesses (50–50) | 180 | 68 | 112 | 3,713 | 1,726 | 1,987 | 32 | 17 | 15 |
| Number of unique videos | 50 | 25 | 25 | 50 | 25 | 25 | 50 | 25 | 25 |
| Observations | 4,236 | 2,131 | 2,105 | 53,925 | 26,566 | 27,359 | 200 | 100 | 100 |
|
| 0.180 | 0.069 | 0.225 | 0.185 | 0.057 | 0.273 | 0.062 | 0.054 | 0.073 |
| Adjusted | 0.170 | 0.056 | 0.215 | 0.184 | 0.057 | 0.272 | 0.048 | 0.025 | 0.044 |
| Residual SE | 0.340 | 0.329 | 0.350 | 0.349 | 0.350 | 0.346 | 0.210 | 0.198 | 0.222 |
| F statistic | 288.686*** | 164.804*** | 169.388*** | 3,687.143*** | 2,150.874*** | 4,525.903*** | 4.337*** | 1.841 | 2.514* |
Linear regressions on participant data include video fixed effects with Eicker–Huber–White SEs clustered at the participant level. * P <0.1; ** P <0.05; *** P <0.01.