William K Diprose1, Nicholas Buist2, Ning Hua3, Quentin Thurier3, George Shand4, Reece Robinson3. 1. Department of Medicine, University of Auckland, Auckland, New Zealand. 2. Department of Emergency Medicine, Whangarei Hospital, Whangarei, New Zealand. 3. Orion Health, Auckland, New Zealand. 4. Clinical Education and Training Unit, Waitematā District Health Board, Auckland, New Zealand.
Abstract
OBJECTIVE: Implementation of machine learning (ML) may be limited by patients' right to "meaningful information about the logic involved" when ML influences healthcare decisions. Given the complexity of healthcare decisions, it is likely that ML outputs will need to be understood and trusted by physicians, and then explained to patients. We therefore investigated the association between physician understanding of ML outputs, their ability to explain these to patients, and their willingness to trust the ML outputs, using various ML explainability methods. MATERIALS AND METHODS: We designed a survey for physicians with a diagnostic dilemma that could be resolved by an ML risk calculator. Physicians were asked to rate their understanding, explainability, and trust in response to 3 different ML outputs. One ML output had no explanation of its logic (the control) and 2 ML outputs used different model-agnostic explainability methods. The relationships among understanding, explainability, and trust were assessed using Cochran-Mantel-Haenszel tests of association. RESULTS: The survey was sent to 1315 physicians, and 170 (13%) provided completed surveys. There were significant associations between physician understanding and explainability (P < .001), between physician understanding and trust (P < .001), and between explainability and trust (P < .001). ML outputs that used model-agnostic explainability methods were preferred by 88% of physicians when compared with the control condition; however, no particular ML explainability method had a greater influence on intended physician behavior. CONCLUSIONS: Physician understanding, explainability, and trust in ML risk calculators are related. Physicians preferred ML outputs accompanied by model-agnostic explanations but the explainability method did not alter intended physician behavior.
OBJECTIVE: Implementation of machine learning (ML) may be limited by patients' right to "meaningful information about the logic involved" when ML influences healthcare decisions. Given the complexity of healthcare decisions, it is likely that ML outputs will need to be understood and trusted by physicians, and then explained to patients. We therefore investigated the association between physician understanding of ML outputs, their ability to explain these to patients, and their willingness to trust the ML outputs, using various ML explainability methods. MATERIALS AND METHODS: We designed a survey for physicians with a diagnostic dilemma that could be resolved by an ML risk calculator. Physicians were asked to rate their understanding, explainability, and trust in response to 3 different ML outputs. One ML output had no explanation of its logic (the control) and 2 ML outputs used different model-agnostic explainability methods. The relationships among understanding, explainability, and trust were assessed using Cochran-Mantel-Haenszel tests of association. RESULTS: The survey was sent to 1315 physicians, and 170 (13%) provided completed surveys. There were significant associations between physician understanding and explainability (P < .001), between physician understanding and trust (P < .001), and between explainability and trust (P < .001). ML outputs that used model-agnostic explainability methods were preferred by 88% of physicians when compared with the control condition; however, no particular ML explainability method had a greater influence on intended physician behavior. CONCLUSIONS: Physician understanding, explainability, and trust in ML risk calculators are related. Physicians preferred ML outputs accompanied by model-agnostic explanations but the explainability method did not alter intended physician behavior.
Authors: Dympna M O'Sullivan; Julie S Doyle; Wojtek J Michalowski; Szymon A Wilk; Ken J Farion; Craig E Kuziemsky Journal: AMIA Annu Symp Proc Date: 2011-10-22
Authors: Scott M Lundberg; Bala Nair; Monica S Vavilala; Mayumi Horibe; Michael J Eisses; Trevor Adams; David E Liston; Daniel King-Wai Low; Shu-Fang Newman; Jerry Kim; Su-In Lee Journal: Nat Biomed Eng Date: 2018-10-10 Impact factor: 25.671
Authors: Sarah A Collins; Kenrick Cato; David Albers; Karen Scott; Peter D Stetson; Suzanne Bakken; David K Vawdrey Journal: Am J Crit Care Date: 2013-07 Impact factor: 2.228
Authors: David J Albers; Matthew Levine; Bruce Gluckman; Henry Ginsberg; George Hripcsak; Lena Mamykina Journal: PLoS Comput Biol Date: 2017-04-27 Impact factor: 4.475
Authors: Seyedeh Neelufar Payrovnaziri; Zhaoyi Chen; Pablo Rengifo-Moreno; Tim Miller; Jiang Bian; Jonathan H Chen; Xiuwen Liu; Zhe He Journal: J Am Med Inform Assoc Date: 2020-07-01 Impact factor: 4.497
Authors: Sunny S Lou; Hanyang Liu; Chenyang Lu; Troy S Wildes; Bruce L Hall; Thomas Kannampallil Journal: Anesthesiology Date: 2022-07-01 Impact factor: 8.986
Authors: Nathaniel Hendrix; Brett Hauber; Christoph I Lee; Aasthaa Bansal; David L Veenstra Journal: J Am Med Inform Assoc Date: 2021-06-12 Impact factor: 4.497
Authors: Morgan Carlile; Brian Hurt; Albert Hsiao; Michael Hogarth; Christopher A Longhurst; Christian Dameff Journal: J Am Coll Emerg Physicians Open Date: 2020-11-05
Authors: Alison L Antes; Sara Burrous; Bryan A Sisk; Matthew J Schuelke; Jason D Keune; James M DuBois Journal: BMC Med Inform Decis Mak Date: 2021-07-20 Impact factor: 2.796