Literature DB >> 32360451

The Potential Dangers of Artificial Intelligence for Radiology and Radiologists.

Linda C Chu1, Anima Anandkumar2, Hoo Chang Shin3, Elliot K Fishman4.   

Abstract

Entities:  

Mesh:

Year:  2020        PMID: 32360451      PMCID: PMC7164850          DOI: 10.1016/j.jacr.2020.04.010

Source DB:  PubMed          Journal:  J Am Coll Radiol        ISSN: 1546-1440            Impact factor:   5.532


× No keyword cloud information.
With the advent of artificial intelligence (AI) across many fields and subspecialties, there are considerable expectations for transformative impact. However, there are also concerns regarding the potential abuse of AI. Many scientists have been worried about the dangers of AI leading to “biased” conclusions, in part because of the enthusiasm of the inventor or overenthusiasm among the general public. Here, though, we consider some scenarios in which people may intend to cause potential errors within data sets of analyzed information, resulting in incorrect conclusions and leading to potential problems with patient care and outcomes. A generative adversarial network (GAN) is a recently developed deep-learning model aimed at creating new images. It simultaneously trains a generator and a discriminator network, which serves to generate artificial images and to discriminate real from artificial images, respectively. We have recently described how GANs can produce artificial images of people and audio content that fool the recipient into believing that they are authentic. As applied to medical imaging, GANs can generate synthetic images that can alter lesion size, location, and transpose abnormalities onto normal examinations (Fig. 1 ) [1]. GANs have the potential to improve image quality, reduce radiation dose, augment data for training algorithms, and perform automated image segmentation [2]. However, there is also the potential for harm if these artificial images infiltrate our health care system by hackers with malicious intent. As proof of principle, Mirsky et al [3] showed that they were able to tamper with CT scans and artificially inject or remove lung cancers on the images. When the radiologists were blinded to the attack, this hack had a 99.2% success rate for cancer injection and a 95.8% success rate for cancer removal. Even when the radiologists were warned about the attack, the success of cancer injection decreased to 70%, but the cancer removal success rate remained high at 90%. This illustrates the sophistication and realistic appearance of such artificial images. These hacks can be targeted against specific patients or can be used as a more general attack on our radiologic data. It is already challenging enough to keep up with the daily clinical volume when the radiology system is running smoothly. Our clinical workflow would be paralyzed if we could not trust the authenticity of the images and must spend extra effort searching for evidence of image tampering in every case.
Fig 1

Examples of images artificially generated using generative adversarial network of brain tumor MR images. First column: T1-weighted images; second column: T1-weighted images with contrast; third column: T2-weighted images; fourth column: fluid-attenuated inversion recovery images. First row: original images with tumor in the right frontal lobe (arrows). Second row: tumor is made 16% larger. Third row: tumor is made 16% smaller. Fourth row: tumor is artificially placed on an otherwise tumor-free brain.

Examples of images artificially generated using generative adversarial network of brain tumor MR images. First column: T1-weighted images; second column: T1-weighted images with contrast; third column: T2-weighted images; fourth column: fluid-attenuated inversion recovery images. First row: original images with tumor in the right frontal lobe (arrows). Second row: tumor is made 16% larger. Third row: tumor is made 16% smaller. Fourth row: tumor is artificially placed on an otherwise tumor-free brain. There are multiple access points within the chain of image acquisition and delivery that can be corrupted by attackers, including the scanner, PACS, server, and workstations [3]. Unfortunately, data security is poorly developed and poorly standardized in radiology. In 2016, Stites and Pianykh [4] performed a scan through the World Wide Web of networked computers and devices and showed that there were 2,774 unprotected radiology or DICOM servers worldwide, most of them located in the United States. To date, there has been no known hack into the radiology system, aside from the research study demonstrating its feasibility [3]. However, the vulnerability is clearly present and may be exploited by hackers. Such threats could affect not only radiology departments but also entire health systems. We have all read articles about security breaches of medical records. There have been almost 3,000 breaches (involving more than 500 medical records) in the United States within the past 10 years. This includes high-profile cases such as the 2015 breach of the Anthem medical insurance company that potentially exposed the medical records of 78 million Americans and led to a $115 million settlement [5]. Hospitals and clinics have been held hostage when their data were corrupted by a third party that demanded payment (ransom) to release the data [5]. In 2017, ransomware WannaCry and NotPetya spread through thousands of institutions worldwide, including many hospitals, and caused $18 billion in damages [5]. Hospitals and clinics have not been the only targets. The city of Baltimore was essentially out of business for a month this past year because of such a ransomware attack. At first glance, all of these situations seem more likely in a movie made for Netflix or HBO. However, the truth is that we must be prepared to deal with such scenarios in the near future. As electronic health records and hospital data become more centralized and more computerized, the dangers only multiply. However, there are several ways to mitigate potential AI-based hacks and attacks. These include clear security guidelines and protocols that are uniform across the globe. As deep-fake technology gets more sophisticated, there is emerging research on AI-driven defense strategies. One example features the training of an AI to detect artificial images by image artifacts induced by GAN [6]. However, AI-driven defense mechanisms have a long way to catch up, as seen in the related problem of defense against adversarial attacks. Recognizing these challenges, the Defense Advanced Research Projects Agency has launched the Media Forensics program to research against deep fakes [7]. Hence, for now, the best defense against deep fakes is based on traditional cybersecurity best practices: secure all stages in the pipeline and enable strong encryption and monitoring tools. In the current coronavirus disease 2019 pandemic, many clinicians and radiologists have turned to working remotely in attempts to “flatten the curve” and slow the spread of disease. In the body imaging division at our institution, currently, approximately half of the radiologists are working remotely. Many of our clinicians are transitioning to telemedicine visits, which adds tremendous stress on our networks. Our IT department has been proactive in setting up a dedicated virtual private network for radiology to ensure that there is sufficient bandwidth for our clinical work. On our few on-site rotations, we practice “social distancing,” and we have suspended our all side-by-side readouts and in-person lectures. We have turned to Zoom (San Jose, California) and other mobile platforms for managing our rapidly changing clinical operations, educating trainees, or simply staying in touch during these uncertain times. The daily meeting participants rose from 10 million daily users in December 2019 to 200 million daily users in March 2020 [8]. Our reliance on Zoom and other mobile platforms has exposed a new vulnerability. There has been a proliferation of “Zoombombing,” in which intruders hijack video calls and paste hate speech and offensive images. Furthermore, additional vulnerabilities in Zoom can allow hackers to gain control of users’ microphones and webcams and steal login credentials. Zoom video meetings did not provide end-to-end encryption as promised, and a large number of Zoom video meeting recordings, many of which contain private information, are left unprotected and viewable on the web. The Federal Bureau of Investigation has issued security warnings about Zoom, and a number of organizations, including SpaceX, Google, New York’s Department of Education, and the US Senate, have banned or discouraged the use of Zoom [8]. The meteoric rise and fall of Zoom is a cautionary tale about the importance of data security. With the development of AI and all its potential wonders in terms of increasing the accuracy of our diagnostic capabilities and potentially improving patient care, we must also be concerned about the potential dark side by bad actors. The sooner organized radiology and organized medicine address these issues with clarity, the more stable and protected the health care system and our patients will be from those intent on creating harm and havoc by abusing AI. The acceleration of data sharing during the current pandemic exposes critical vulnerabilities in data security. It reminds us of the pervasive threat that bad actors can and will exploit any technology for their selfish gains. Doing nothing is not a viable strategy, but acting in a concerted effort will lead us to the protection we need and is important as we push AI development over the next several years.
  3 in total

1.  How Secure Is Your Radiology Department? Mapping Digital Radiology Adoption and Security Worldwide.

Authors:  Mark Stites; Oleg S Pianykh
Journal:  AJR Am J Roentgenol       Date:  2016-03-02       Impact factor: 3.959

Review 2.  Creating Artificial Images for Radiology Applications Using Generative Adversarial Networks (GANs) - A Systematic Review.

Authors:  Vera Sorin; Yiftach Barash; Eli Konen; Eyal Klang
Journal:  Acad Radiol       Date:  2020-02-05       Impact factor: 3.173

Review 3.  DICOM Images Have Been Hacked! Now What?

Authors:  Benoit Desjardins; Yisroel Mirsky; Markel Picado Ortiz; Zeev Glozman; Lawrence Tarbox; Robert Horn; Steven C Horii
Journal:  AJR Am J Roentgenol       Date:  2019-11-26       Impact factor: 3.959

  3 in total
  2 in total

Review 1.  Applications of Generative Adversarial Networks (GANs) in Positron Emission Tomography (PET) imaging: A review.

Authors:  Ioannis D Apostolopoulos; Nikolaos D Papathanasiou; Dimitris J Apostolopoulos; George S Panayiotakis
Journal:  Eur J Nucl Med Mol Imaging       Date:  2022-04-22       Impact factor: 10.057

2.  The evaluation of the reduction of radiation dose via deep learning-based reconstruction for cadaveric human lung CT images.

Authors:  Tomo Miyata; Masahiro Yanagawa; Noriko Kikuchi; Kazuki Yamagata; Yukihisa Sato; Yuriko Yoshida; Mitsuko Tsubamoto; Noriyuki Tomiyama
Journal:  Sci Rep       Date:  2022-07-20       Impact factor: 4.996

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.