Jooyoung Chang1, Jinho Lee2, Ahnul Ha3, Young Soo Han3, Eunoo Bak3, Seulggie Choi1, Jae Moon Yun4, Uk Kang5, Il Hyung Shin6, Joo Young Shin7, Taehoon Ko8, Ye Seul Bae9, Baek-Lok Oh3, Ki Ho Park10, Sang Min Park11. 1. Department of Biomedical Sciences, Seoul National University Graduate School, Seoul, Republic of Korea. 2. Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Republic of Korea; Department of Ophthalmology, Hallym University Chuncheon Sacred Heart Hospital, Chuncheon, Republic of Korea. 3. Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Republic of Korea; Department of Ophthalmology, Seoul National University Hospital, Seoul, Republic of Korea. 4. Department of Family Medicine, Seoul National University Hospital, Seoul, Republic of Korea. 5. InTheSmart Co., Ltd., Seoul, Republic of Korea; Biomedical Research Institute, Seoul National University Hospital, Seoul, Republic of Korea. 6. InTheSmart Co., Ltd., Seoul, Republic of Korea. 7. Department of Ophthalmology, Seoul Metropolitan Government Seoul National University Boramae Medical Center, Seoul, Republic of Korea. 8. Office of Hospital Information, Seoul National University Hospital, Seoul, Republic of Korea. 9. Department of Family Medicine, Seoul National University Hospital, Seoul, Republic of Korea; Office of Hospital Information, Seoul National University Hospital, Seoul, Republic of Korea. 10. Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Republic of Korea; Department of Ophthalmology, Seoul National University Hospital, Seoul, Republic of Korea. Electronic address: kihopark@snu.ac.kr. 11. Department of Biomedical Sciences, Seoul National University Graduate School, Seoul, Republic of Korea; Department of Family Medicine, Seoul National University Hospital, Seoul, Republic of Korea. Electronic address: smpark.snuh@gmail.com.
Abstract
PURPOSE: To illustrate what is inside the so-called black box of deep learning models (DLMs) so that clinicians can have greater confidence in the conclusions of artificial intelligence by evaluating adversarial explanation on its ability to explain the rationale of DLM decisions for glaucoma and glaucoma-related findings. Adversarial explanation generates adversarial examples (AEs), or images that have been changed to gain or lose pathologic characteristic-specific traits, to explain the DLM's rationale. DESIGN: Evaluation of explanation methods for DLMs. PARTICIPANTS: Health screening participants (n = 1653) at the Seoul National University Hospital Health Promotion Center, Seoul, Republic of Korea. METHODS: We trained DLMs for referable glaucoma (RG), increased cup-to-disc ratio (ICDR), disc rim narrowing (DRN), and retinal nerve fiber layer defect (RNFLD) using 6430 retinal fundus images. Surveys consisting of explanations using AE and gradient-weighted class activation mapping (GradCAM), a conventional heatmap-based explanation method, were generated for 400 pathologic and healthy patient eyes. For each method, board-trained glaucoma specialists rated location explainability, the ability to pinpoint decision-relevant areas in the image, and rationale explainability, the ability to inform the user on the model's reasoning for the decision based on pathologic features. Scores were compared by paired Wilcoxon signed-rank test. MAIN OUTCOME MEASURES: Area under the receiver operating characteristic curve (AUC), sensitivities, and specificities of DLMs; visualization of clinical pathologic changes of AEs; and survey scores for locational and rationale explainability. RESULTS: The AUCs were 0.90, 0.99, 0.95, and 0.79 and sensitivities were 0.79, 1.00, 0.82, and 0.55 at 0.90 specificity for RG, ICDR, DRN, and RNFLD DLMs, respectively. Generated AEs showed valid clinical feature changes, and survey results for location explainability were 3.94 ± 1.33 and 2.55 ± 1.24 using AEs and GradCAMs, respectively, of a possible maximum score of 5 points. The scores for rationale explainability were 3.97 ± 1.31 and 2.10 ± 1.25 for AEs and GradCAM, respectively. Adversarial example provided significantly better explainability than GradCAM. CONCLUSIONS: Adversarial explanation increased the explainability over GradCAM, a conventional heatmap-based explanation method. Adversarial explanation may help medical professionals understand more clearly the rationale of DLMs when using them for clinical decisions.
PURPOSE: To illustrate what is inside the so-called black box of deep learning models (DLMs) so that clinicians can have greater confidence in the conclusions of artificial intelligence by evaluating adversarial explanation on its ability to explain the rationale of DLM decisions for glaucoma and glaucoma-related findings. Adversarial explanation generates adversarial examples (AEs), or images that have been changed to gain or lose pathologic characteristic-specific traits, to explain the DLM's rationale. DESIGN: Evaluation of explanation methods for DLMs. PARTICIPANTS: Health screening participants (n = 1653) at the Seoul National University Hospital Health Promotion Center, Seoul, Republic of Korea. METHODS: We trained DLMs for referable glaucoma (RG), increased cup-to-disc ratio (ICDR), disc rim narrowing (DRN), and retinal nerve fiber layer defect (RNFLD) using 6430 retinal fundus images. Surveys consisting of explanations using AE and gradient-weighted class activation mapping (GradCAM), a conventional heatmap-based explanation method, were generated for 400 pathologic and healthy patient eyes. For each method, board-trained glaucoma specialists rated location explainability, the ability to pinpoint decision-relevant areas in the image, and rationale explainability, the ability to inform the user on the model's reasoning for the decision based on pathologic features. Scores were compared by paired Wilcoxon signed-rank test. MAIN OUTCOME MEASURES: Area under the receiver operating characteristic curve (AUC), sensitivities, and specificities of DLMs; visualization of clinical pathologic changes of AEs; and survey scores for locational and rationale explainability. RESULTS: The AUCs were 0.90, 0.99, 0.95, and 0.79 and sensitivities were 0.79, 1.00, 0.82, and 0.55 at 0.90 specificity for RG, ICDR, DRN, and RNFLD DLMs, respectively. Generated AEs showed valid clinical feature changes, and survey results for location explainability were 3.94 ± 1.33 and 2.55 ± 1.24 using AEs and GradCAMs, respectively, of a possible maximum score of 5 points. The scores for rationale explainability were 3.97 ± 1.31 and 2.10 ± 1.25 for AEs and GradCAM, respectively. Adversarial example provided significantly better explainability than GradCAM. CONCLUSIONS: Adversarial explanation increased the explainability over GradCAM, a conventional heatmap-based explanation method. Adversarial explanation may help medical professionals understand more clearly the rationale of DLMs when using them for clinical decisions.
Authors: Xiaoling Huang; Kai Jin; Jiazhu Zhu; Ying Xue; Ke Si; Chun Zhang; Sukun Meng; Wei Gong; Juan Ye Journal: Front Med (Lausanne) Date: 2022-03-17
Authors: Daniele Ravi; Stefano B Blumberg; Silvia Ingala; Frederik Barkhof; Daniel C Alexander; Neil P Oxtoby Journal: Med Image Anal Date: 2021-10-14 Impact factor: 8.545