Literature DB >> 30399149

Machine learning in medicine: Addressing ethical challenges.

Effy Vayena1, Alessandro Blasimme1, I Glenn Cohen2.   

Abstract

Effy Vayena and colleagues argue that machine learning in medicine must offer data protection, algorithmic transparency, and accountability to earn the trust of patients and clinicians.

Entities:  

Mesh:

Year:  2018        PMID: 30399149      PMCID: PMC6219763          DOI: 10.1371/journal.pmed.1002689

Source DB:  PubMed          Journal:  PLoS Med        ISSN: 1549-1277            Impact factor:   11.069


A recent United Kingdom survey reports that 63% of the adult population is uncomfortable with allowing personal data to be used to improve healthcare and is unfavorable to artificial intelligence (AI) systems replacing doctors and nurses in tasks they usually perform [1]. Another study, conducted in Germany, found that medical students—the doctors of tomorrow—overwhelmingly buy into the promise of AI to improve medicine (83%) but are more skeptical that it will establish conclusive diagnoses in, for instance, imaging exams (56% disagree) [2]. When asked about the prospects of AI, United States decision-makers at healthcare organizations are confident that it will improve medicine, but roughly half of them think it will produce fatal errors, will not work properly, and will not meet currently hyped expectations [3]. These survey data resonate to the ethical and regulatory challenges that surround AI in healthcare, particularly privacy, data fairness, accountability, transparency, and liability. Successfully addressing these will foster the future of machine learning in medicine (MLm) and its positive impact on healthcare. Ethical and regulatory concerns about MLm can be grouped into three broad categories according to how and where they emerge, namely, around the sources of data needed for MLm and the approaches used in the development and deployment of MLm in clinical practice (for a hypothetical case study, see Fig 1).
Fig 1

Imagine a medical software company developing a machine learning–based device.

The device performs fully automated analysis of histopathology slides from cancer patients and predicts genetic mutations in tumors solely based on these images. This inferred genetic information can be used either for prognostic purposes or to detect an indication for a targeted therapy. Users will not know which features of the images the algorithm associates with mutated genes or the biological explanation for these associations. The selling propositions of the device are that it can infer valuable genetic information early in the diagnostic process and be used in contexts in which genetic testing is not available by analyzing images shared by pathologists on a cloud-based platform. MLm, machine learning in medicine.

Imagine a medical software company developing a machine learning–based device.

The device performs fully automated analysis of histopathology slides from cancer patients and predicts genetic mutations in tumors solely based on these images. This inferred genetic information can be used either for prognostic purposes or to detect an indication for a targeted therapy. Users will not know which features of the images the algorithm associates with mutated genes or the biological explanation for these associations. The selling propositions of the device are that it can infer valuable genetic information early in the diagnostic process and be used in contexts in which genetic testing is not available by analyzing images shared by pathologists on a cloud-based platform. MLm, machine learning in medicine.

Data sourcing for MLm must adhere to data protection and privacy requirements

MLm algorithms use data that are subject to privacy protections, requiring that developers pay close attention to ethical and regulatory restrictions at each stage of data processing. Data provenance and consent for use and reuse are of particular importance [4,5], especially for MLm that requires considerable amounts and large varieties of data. It is very likely that such disparate data will have different conditions of use and/or be bound by different legal protections. A prominent example is the newly enacted European General Data Protection Regulation (GDPR), which sets out specific informed consent requirements for data uses and grants data subjects several rights that must be respected by those processing their data [6]. Moreover, this law applies to data from residents of the European Union (EU) irrespective of where the data are processed. Data that are used to train algorithms must have the necessary use authorizations, but determining which data uses are permitted for a given purpose is not an easy feat. This will also depend on data type, jurisdiction, purpose of use, and oversight models. Ethical and regulatory concerns around health data have been debated for some time, and regulation such as GDPR is emblematic of legislative motivation to address novel digital challenges. However, different jurisdictions have reached and can be expected to reach different conclusions. Compare, for example, the GDPR with the Health Insurance Portability and Accountability Act [HIPAA] in the US, which focuses on healthcare data from patient records but does not cover other sources of health data such as health data generated outside of covered entities and business associates, e.g., data generated by life insurance companies or by a blood glucose–monitoring smartphone app [7].

MLm development should be committed to fairness

The computer science adage goes, “garbage in, garbage out.” This is especially true for MLm, since the data sets on which MLm models are trained and validated are essential in ensuring the ethical use of predictive algorithms. Poorly representative training data sets can introduce biases into MLm-trained algorithms. “Bias” is a fraught term, with at least two archetypes common in medical data. First are cases in which the data sources themselves do not reflect true epidemiology within a given demographic, as for instance in population data biased by the entrenched overdiagnosis of schizophrenia in African Americans [8]. Second are cases in which an algorithm is trained on a data set that does not contain enough members of a given demographic—for instance, an algorithm trained mostly on data from older white men. Such an algorithm would make poor predictions, for example, among younger black women. If algorithms trained on data sets with these characteristics are adopted in healthcare, they have the potential to exacerbate health disparities [9]. To avoid this pitfall, scientific societies and regulatory agencies must develop best practices for recognizing and minimizing the downstream effects of biased training data sets, while bodies such as institutional review boards, ethics review committees, and health technology assessment organizations should check for compliance with such standards. As an example of one opportunity to reduce the risks of biased algorithms, the US Food and Drug Administration (FDA), in the context of its Digital Health Innovation Action Plan, has started a precertification pilot program to assess medical software under development based on five excellence criteria, including quality [10]. To avert the harms of biased training sets, the FDA quality criterion—and other regulatory criteria like it—could be expanded to cover the risk of bias in training data sets. This need is being addressed in some circles; for instance, the recent American Medical Association (AMA) policy recommendations on AI (in this case, “augmented intelligence”) promote the identification and minimization of biases in data [11]. A different approach, exemplified by the All of Us precision medicine research cohort in the US, is to fund the development of more representative data sets that can be used for training and validation [12]. In addition to regulatory protections against algorithmic bias, modernized regulatory approval—in the form of ad hoc guidance—is needed to maintain the safety and efficacy of MLm-based algorithms. These algorithms can in principle improve continuously with each use, resulting in real-time “updates” that cannot possibly be tested individually in clinical trials or assessed according to the typical timetable of a healthcare regulator. Regulators must develop standard procedures including effective postmarket monitoring mechanisms by which developers can transparently document the evolution of their MLm software.

The deployment of MLm should satisfy transparency standards

Perhaps the MLm raising the most difficult ethical and legal questions—and the greatest challenge to current modes of regulation—is represented by noninterpretable, so-called black-box algorithms, the inner logic of which remains hidden even to their developers. This lack of transparency can preclude the mechanistic interpretation of MLm-based assessments and, in turn, reduce their trustworthiness. Moreover, the disclosure of basic yet meaningful details about medical treatment to patients—a fundamental tenet of medical ethics—requires that the doctors themselves grasp at least the fundamental inner workings of the devices they use. Therefore, for MLm to be ethical, developers must communicate to their end users—doctors—the general logic behind MLm-based decisions. Some degree of explainability may also be required to justify the clinical validation of MLm in prospective studies and randomized clinical trials. In the case of fully automated medical decisions, the level of risk associated with the procedure may determine whether and how information should be provided to patients about the presence of MLm-based technologies employed to guide their care. Communicating with patients about the use of MLm technologies may increase their trust and acceptance, which the survey data discussed above suggest is an ongoing concern. As more diagnostic and therapeutic interventions become based on MLm, the autonomy of patients in decisional processes about their health and the possibility of shared decision-making may be undermined. This would happen, for instance, if reliance on automated decision-making tools reduces the opportunity of meaningful dialogue between healthcare providers and patients or if payers consider MLm recommendations as a precondition for reimbursement and refuse to cover treatments when the MLm recommends against them. Given the importance of personal contact and interaction between patients, healthcare professionals, and caregivers, it is key for healthcare providers to embed MLm in supportive and empowering delivery systems. Third-party auditing may also prove necessary, and such ethical integration may become a precondition for accreditation. Special attention should be paid to the AI-enabled consumer informatics such as smartphone apps and wearable devices. Rapid progress in machine learning could result in a proliferation of health-related consumer products for which quality standards and accreditation systems should be developed. Finally, the allocation and grounds for liability for adverse events related to the use of MLm will need to be clarified. Without a thorough ability to predict how various kinds of MLm will expose hospitals, physicians, and developers to liability, it will be hard to achieve widespread adoption [13,14].

Conclusions

The clinical use of MLm may transform existing modes of healthcare delivery. MLm will be used in the clinical setting by healthcare professionals, be embedded in smart devices through the internet of things, and be used by patients themselves beyond the clinical setting for disease self-management of chronic conditions. The exponential growth of investment in MLm signals that research is accelerating, and more products may soon be targeting market entry. To merit the trust of patients and adoption by providers, MLm must fully align with data protection requirements, minimize the effects of bias, be effectively regulated, and achieve transparency. Addressing such ethical and regulatory issues as soon as possible is essential for avoiding unnecessary risks and pitfalls that will hinder further progress of MLm.
  8 in total

Review 1.  Health disparities and health equity: concepts and measurement.

Authors:  Paula Braveman
Journal:  Annu Rev Public Health       Date:  2006       Impact factor: 21.981

2.  HIPAA and Protecting Health Information in the 21st Century.

Authors:  I Glenn Cohen; Michelle M Mello
Journal:  JAMA       Date:  2018-07-17       Impact factor: 56.272

3.  Medical students' attitude towards artificial intelligence: a multicentre survey.

Authors:  D Pinto Dos Santos; D Giese; S Brodehl; S H Chon; W Staab; R Kleinert; D Maintz; B Baeßler
Journal:  Eur Radiol       Date:  2018-07-06       Impact factor: 5.315

Review 4.  The influence of racial factors on psychiatric diagnosis: a review and suggestions for research.

Authors:  H W Neighbors; J S Jackson; L Campbell; D Williams
Journal:  Community Ment Health J       Date:  1989

5.  What does the GDPR mean for the medical community?

Authors:  Becky McCall
Journal:  Lancet       Date:  2018-03-31       Impact factor: 79.321

6.  Health Research with Big Data: Time for Systemic Oversight.

Authors:  Effy Vayena; Alessandro Blasimme
Journal:  J Law Med Ethics       Date:  2018-03-27       Impact factor: 1.718

7.  Regulating Black-Box Medicine.

Authors:  W Nicholson Price
Journal:  Mich Law Rev       Date:  2017

8.  Biomedical Big Data: New Models of Control Over Access, Use and Governance.

Authors:  Effy Vayena; Alessandro Blasimme
Journal:  J Bioeth Inq       Date:  2017-10-05       Impact factor: 1.352

  8 in total
  72 in total

Review 1.  Artificial Intelligence in radiotherapy: state of the art and future directions.

Authors:  Giulio Francolini; Isacco Desideri; Giulia Stocchi; Viola Salvestrini; Lucia Pia Ciccone; Pietro Garlatti; Mauro Loi; Lorenzo Livi
Journal:  Med Oncol       Date:  2020-04-22       Impact factor: 3.064

2.  Reporting of data analysis methods in psychiatric journals: Trends from 1996 to 2018.

Authors:  Pentti Nieminen; Jasleen Kaur
Journal:  Int J Methods Psychiatr Res       Date:  2019-05-27       Impact factor: 4.035

Review 3.  Looking beyond the hype: Applied AI and machine learning in translational medicine.

Authors:  Tzen S Toh; Frank Dondelinger; Dennis Wang
Journal:  EBioMedicine       Date:  2019-08-26       Impact factor: 8.143

4.  Fostering a Healthy AI Ecosystem for Radiology: Conclusions of the 2018 RSNA Summit on AI in Radiology.

Authors:  Falgun H Chokshi; Adam E Flanders; Luciano M Prevedello; Curtis P Langlotz
Journal:  Radiol Artif Intell       Date:  2019-03-27

5.  Machine Learning in Oncology: Methods, Applications, and Challenges.

Authors:  Dimitris Bertsimas; Holly Wiberg
Journal:  JCO Clin Cancer Inform       Date:  2020-10

6.  A governance model for the application of AI in health care.

Authors:  Sandeep Reddy; Sonia Allan; Simon Coghlan; Paul Cooper
Journal:  J Am Med Inform Assoc       Date:  2020-03-01       Impact factor: 4.497

7.  Between-laboratory reproducibility of time-lapse embryo selection using qualitative and quantitative parameters: a systematic review and meta-analysis.

Authors:  Yanhe Liu; Fang Qi; Phillip Matson; Dean E Morbeck; Ben W Mol; Sai Zhao; Masoud Afnan
Journal:  J Assist Reprod Genet       Date:  2020-05-02       Impact factor: 3.412

8.  Latent bias and the implementation of artificial intelligence in medicine.

Authors:  Matthew DeCamp; Charlotta Lindvall
Journal:  J Am Med Inform Assoc       Date:  2020-12-09       Impact factor: 4.497

9.  The Challenges of Implementing Artificial Intelligence into Surgical Practice.

Authors:  Isaac Tranter-Entwistle; Holly Wang; Kenny Daly; Scott Maxwell; Saxon Connor
Journal:  World J Surg       Date:  2020-10-13       Impact factor: 3.352

Review 10.  Machine learning for sperm selection.

Authors:  Jae Bem You; Christopher McCallum; Yihe Wang; Jason Riordon; Reza Nosrati; David Sinton
Journal:  Nat Rev Urol       Date:  2021-05-17       Impact factor: 14.432

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.