| Literature DB >> 34907229 |
Qianwei Zhou1,2,3, Margarita Zuley1,4, Yuan Guo1,5, Lu Yang1,6, Bronwyn Nair1,4, Adrienne Vargo1,4, Suzanne Ghannam1,4, Dooman Arefan1, Shandong Wu7,8,9,10.
Abstract
While active efforts are advancing medical artificial intelligence (AI) model development and clinical translation, safety issues of the AI models emerge, but little research has been done. We perform a study to investigate the behaviors of an AI diagnosis model under adversarial images generated by Generative Adversarial Network (GAN) models and to evaluate the effects on human experts when visually identifying potential adversarial images. Our GAN model makes intentional modifications to the diagnosis-sensitive contents of mammogram images in deep learning-based computer-aided diagnosis (CAD) of breast cancer. In our experiments the adversarial samples fool the AI-CAD model to output a wrong diagnosis on 69.1% of the cases that are initially correctly classified by the AI-CAD model. Five breast imaging radiologists visually identify 29%-71% of the adversarial samples. Our study suggests an imperative need for continuing research on medical AI model's safety issues and for developing potential defensive solutions against adversarial attacks.Entities:
Mesh:
Year: 2021 PMID: 34907229 PMCID: PMC8671500 DOI: 10.1038/s41467-021-27577-x
Source DB: PubMed Journal: Nat Commun ISSN: 2041-1723 Impact factor: 14.919
Fig. 1An overview of our study design.
An AI-CAD model was first learned and then tested on the adversarial images generated by the GAN model which aimed to make modifications to the diagnosis-sensitive contents of images (by inserting or removing cancerous tissue). The reader study examined human experts’ capabilities to visually recognize the GAN-generated adversarial images.
Classification effects of the AI-CAD model on the test data (74 real positive samples and 364 real negative samples, and their corresponding GAN-generated fake images) at two different resolutions.
| Image resolution | AUC on real images (74 positive and 364 negative samples) | AUC on the corresponding GAN-generated fake counterparts (label flipped) | Classes of the real images | Percentage of correctly classified real images | Percentage of the fake counterparts (of the correctly classified real images) that fooled the AI-CAD model |
|---|---|---|---|---|---|
| 1728 × 1408 | 0.82 | 0.94 | Positive | 59.5% (44/74) | 95.5% (42/44) |
| Negative | 87.6% (319/364) | 65.5% (209/319) | |||
| 1024 × 832 | 0.82 | 0.79 | Positive | 58.1% (43/74) | 88.4% (38/43) |
| Negative | 83.2% (303/364) | 66.7% (202/303) |
Fig. 2Examples of the images shown to the readers in the educational intervention.
Each case consisted of a real image, the synthetic adversarial sample generated by the GAN model, and the difference calculated by the subtraction (real - synthetic) between the two images. Note that the arrows appeared in the second row were not part of the images shown for the educational purpose; they were provided here to indicate important changes made by the GAN models to the images.
Number of images used in different sessions in the human reader study (each sample in Session 4 consisted of one real image, its synthetic counterpart, and the difference/subtraction image).
| Session # | Real images | GAN-generated fake images | Total | ||
|---|---|---|---|---|---|
| Positive samples | Negative samples | Positive-looking counterparts | Negative-looking counterparts | ||
| 1 | 51 | 49 | – | – | 100 |
| 2 | 49 | 51 | – | – | 100 |
| 3 | 36 | 179 | 183 | 38 | 436 |
| 4 | 50 | 50 | 50 | 50 | 100 |
| 5 | 39 | 184 | 180 | 37 | 440 |
Accuracies of correctly identifying real/fake images at each session of the human reader study. Real images are the original images while fake images are the GAN-generated adversarial images.
| Reader ID (number of years of clinical experience) | Mixed real and fake images | Real images (positive & negative cases) | Fake images (positive & negative cases) | Real images (positive cases) | Real images (negative cases) | Fake images (positive-looking images) | Fake images (negative-looking images) | Not sure an image is real or fake |
|---|---|---|---|---|---|---|---|---|
| High-resolution Images (1728 × 1408) | ||||||||
| Session 2 | ||||||||
| Reader 1 (14) | – | 0.89 | – | 0.88 | 0.90 | – | – | 0.05 |
| Reader 2 (13) | – | 0.60 | – | 0.59 | 0.61 | – | – | 0.00 |
| Reader 3 (12) | – | 0.92 | – | 0.96 | 0.88 | – | – | 0.03 |
| Reader 4 (7) | – | 0.60 | – | 0.67 | 0.53 | – | – | 0.00 |
| Reader 5 (<1) | – | 0.53 | – | 0.55 | 0.51 | – | – | 0.02 |
| Session 3 | ||||||||
| Reader 1 (14) | 0.61 | 0.94 | 0.29 | 0.94 | 0.94 | 0.3 | 0.24 | 0.02 |
| Reader 2 (13) | 0.58 | 0.65 | 0.51 | 0.64 | 0.65 | 0.53 | 0.39 | 0.00 |
| Reader 3 (12) | 0.67 | 1.00 | 0.35 | 1.00 | 0.99 | 0.35 | 0.34 | 0.03 |
| Reader 4 (7) | 0.82 | 0.94 | 0.71 | 0.97 | 0.93 | 0.71 | 0.68 | 0.00 |
| Reader 5 (<1) | 0.53 | 0.61 | 0.44 | 0.61 | 0.61 | 0.43 | 0.5 | 0.01 |
| Session 5 | ||||||||
| Reader 1 (14) | 0.58 | 0.85 | 0.30 | 0.72 | 0.88 | 0.32 | 0.22 | 0.06 |
| Reader 2 (13) | 0.57 | 0.49 | 0.65 | 0.31 | 0.53 | 0.72 | 0.32 | 0.00 |
| Reader 3 (12) | 0.75 | 0.97 | 0.51 | 0.97 | 0.97 | 0.49 | 0.59 | 0.03 |
| Reader 4 (7) | 0.82 | 0.95 | 0.70 | 1.00 | 0.93 | 0.67 | 0.84 | 0.00 |
| Reader 5 (<1) | 0.46 | 0.65 | 0.27 | 0.69 | 0.65 | 0.28 | 0.22 | 0.02 |
| Low-resolution Images (1024 × 832) | ||||||||
| Session 2 | ||||||||
| Reader 1 (14) | – | 0.86 | – | 0.80 | 0.92 | – | – | 0.03 |
| Reader 2 (13) | – | 0.71 | – | 0.71 | 0.69 | – | – | 0.00 |
| Session 3 | ||||||||
| Reader 1 (14) | 0.77 | 0.93 | 0.61 | 0.89 | 0.94 | 0.66 | 0.34 | 0.02 |
| Reader 2 (13) | 0.61 | 0.92 | 0.31 | 0.86 | 0.93 | 0.34 | 0.11 | 0.00 |
| Session 5 | ||||||||
| Reader 1 (14) | 0.76 | 0.88 | 0.64 | 0.90 | 0.88 | 0.69 | 0.38 | 0.00 |
| Reader 2 (13) | 0.59 | 0.71 | 0.47 | 0.59 | 0.73 | 0.52 | 0.27 | 0.06 |
Time spent in each session (unit: minute) of the reader study.
| Reader | Session 1 | Session 2 | Session 3 | Session 4 | Session 5 |
|---|---|---|---|---|---|
| High-resolution Images | |||||
| Reader 1 | 10 | 11 | 61 | 22 | 80 |
| Reader 2 | 32 | 25 | 100 | 30 | 130 |
| Reader 3 | 14 | 16 | 57 | 19 | 65 |
| Reader 4 | 9 | 17 | 51 | 31 | 46 |
| Reader 5 | 38 | 34 | 136 | 32 | 128 |
| Low-resolution Images | |||||
| Reader 1 | 5 | 20 | 52 | 15 | 80 |
| Reader 2 | 8 | 20 | 60 | 10 | 120 |
Fig. 3Network design of the models used in our study.
The structure of the AI-CAD classifier (a) and the GAN generator (b).
Fig. 4The GAN model training procedures for generating negative-looking fake/adversarial images.
The procedures (not depicted here) were similar for generating positive-looking fake/adversarial images.