Literature DB >> 27809908

Cognitive biases associated with medical decisions: a systematic review.

Gustavo Saposnik1,2,3,4, Donald Redelmeier5, Christian C Ruff6, Philippe N Tobler6.   

Abstract

BACKGROUND: Cognitive biases and personality traits (aversion to risk or ambiguity) may lead to diagnostic inaccuracies and medical errors resulting in mismanagement or inadequate utilization of resources. We conducted a systematic review with four objectives: 1) to identify the most common cognitive biases, 2) to evaluate the influence of cognitive biases on diagnostic accuracy or management errors, 3) to determine their impact on patient outcomes, and 4) to identify literature gaps.
METHODS: We searched MEDLINE and the Cochrane Library databases for relevant articles on cognitive biases from 1980 to May 2015. We included studies conducted in physicians that evaluated at least one cognitive factor using case-vignettes or real scenarios and reported an associated outcome written in English. Data quality was assessed by the Newcastle-Ottawa scale. Among 114 publications, 20 studies comprising 6810 physicians met the inclusion criteria. Nineteen cognitive biases were identified.
RESULTS: All studies found at least one cognitive bias or personality trait to affect physicians. Overconfidence, lower tolerance to risk, the anchoring effect, and information and availability biases were associated with diagnostic inaccuracies in 36.5 to 77 % of case-scenarios. Five out of seven (71.4 %) studies showed an association between cognitive biases and therapeutic or management errors. Of two (10 %) studies evaluating the impact of cognitive biases or personality traits on patient outcomes, only one showed that higher tolerance to ambiguity was associated with increased medical complications (9.7 % vs 6.5 %; p = .004). Most studies (60 %) targeted cognitive biases in diagnostic tasks, fewer focused on treatment or management (35 %) and on prognosis (10 %). Literature gaps include potentially relevant biases (e.g. aggregate bias, feedback sanction, hindsight bias) not investigated in the included studies. Moreover, only five (25 %) studies used clinical guidelines as the framework to determine diagnostic or treatment errors. Most studies (n = 12, 60 %) were classified as low quality.
CONCLUSIONS: Overconfidence, the anchoring effect, information and availability bias, and tolerance to risk may be associated with diagnostic inaccuracies or suboptimal management. More comprehensive studies are needed to determine the prevalence of cognitive biases and personality traits and their potential impact on physicians' decisions, medical errors, and patient outcomes.

Entities:  

Keywords:  Case-scenarios; Cognition; Cognitive bias; Decision making; Personality traits; Physicians; Systematic review

Mesh:

Year:  2016        PMID: 27809908      PMCID: PMC5093937          DOI: 10.1186/s12911-016-0377-1

Source DB:  PubMed          Journal:  BMC Med Inform Decis Mak        ISSN: 1472-6947            Impact factor:   2.796


Background

Medical errors occur in 1.7-6.5 % of all hospital admissions causing up to 100,000 unnecessary deaths each year, and perhaps one million in excess injuries in the USA [1, 2]. In 2008, medical errors cost the USA $19.5 billion [3]. The incremental cost associated with the average event was about US$ 4685 and an increased length of stay of about 4.6 days. The ultimate consequences of medical errors include avoidable hospitalizations, medication underuse and overuse, and wasted resources that may lead to patients’ harm [4, 5]. Kahneman and Tversky introduced a dual-system theoretical framework to explain judgments, decisions under uncertainty, and cognitive biases. System 1 refers to an automatic, intuitive, unconscious, fast, and effortless or routine mechanism to make most common decisions (Fig. 1). Conversely, system 2 makes deliberate decisions, which are non-programmed, conscious, usually slow and effortful [6]. It has been suggested that most cognitive biases are likely due to the overuse of system 1 or when system 1 overrides system 2 [7-9]. In this framework, techniques that enhance system 2 could counteract these biases and thereby improve diagnostic accuracy and decrease management errors.
Fig. 1

A model for diagnostic reasoning based on dual-process theory (from Ely et al. with permission).[9] System 1 thinking can be influenced by multiple factors, many of them subconscious (emotional polarization toward the patient, recent experience with the diagnosis being considered, specific cognitive or affective biases), and is therefore represented with multiple channels, whereas system 2 processes are, in a given instance, single-channeled and linear. System 2 overrides system 1 (executive override) when physicians take a time-out to reflect on their thinking, possibly with the help of checklists. In contrast, system 1 may irrationally override system 2 when physicians insist on going their own way (e.g., ignoring evidence-based clinical decision rules that can usually outperform them). Notes: Dysrationalia denotes the inability to think rationally despite adequate intelligence. “Calibration” denotes the degree to which the perceived and actual diagnostic accuracy correspond

A model for diagnostic reasoning based on dual-process theory (from Ely et al. with permission).[9] System 1 thinking can be influenced by multiple factors, many of them subconscious (emotional polarization toward the patient, recent experience with the diagnosis being considered, specific cognitive or affective biases), and is therefore represented with multiple channels, whereas system 2 processes are, in a given instance, single-channeled and linear. System 2 overrides system 1 (executive override) when physicians take a time-out to reflect on their thinking, possibly with the help of checklists. In contrast, system 1 may irrationally override system 2 when physicians insist on going their own way (e.g., ignoring evidence-based clinical decision rules that can usually outperform them). Notes: Dysrationalia denotes the inability to think rationally despite adequate intelligence. “Calibration” denotes the degree to which the perceived and actual diagnostic accuracy correspond Concerns about cognitive biases are not unique to medicine. Previous studies showed the influence of cognitive biases on decisions inducing errors in other fields (e.g., aeronautic industry, factory production) [10, 11]. For example, a study investigating failures and accidents identified that over 90 % of air traffic control system errors, 82 % of production errors in an unnamed company, and 50–70 % of all electronic equipment failures were partly or wholly due to human cognitive factors [10]. Psychological assessments and quality assessment tools (e.g. Six Sigma) have been applied in many sectors to reduce errors and improve quality [12-15]. The health sector shares commonalities with industrial sectors including vulnerability to human errors [11, 14]. Therefore, a better understanding of the available evidence on cognitive biases influencing medical decisions is crucial. Such an understanding is particularly needed for physicians, as their errors can be fatal and very costly. Moreover, such an understanding could also be useful to inform learning strategies to improve clinical performance and patient outcomes, whereas literature gaps could be useful to inform future research. In the last three decades, we learned about the importance of patient- and hospital-level factors associated with medical errors. For example, standardized approaches (e.g. Advanced Trauma Life Support, ABCs for cardiopulmonary resuscitation) at the health system level lead to better outcomes by decreasing medical errors [16, 17]. However, physician-level factors were largely ignored as reflected by reports from scientific organizations [18-20]. It was not until the 1970s that cognitive biases were initially recognized to affect individual physicians’ performance in daily medical decisions [6, 21–24]. Despite these efforts, little is known about the influence of cognitive biases and personality traits on physicians’ decisions that lead to diagnostic inaccuracies, medical errors or impact on patient outcomes. While a recent review on cognitive biases and heuristics suggested that general medical personnel is prone to show cognitive biases, it did not answer the question whether these biases actually relate to the number of medical errors in physicians [25]. In the present (primarily narrative) systematic review, we therefore reviewed the literature reporting the existing evidence on the relation between cognitive biases affecting physicians and medical decisions. Under the concept of cognitive biases, we also included personality traits (e.g. aversion to risk or ambiguity) that may systematically affect physicians’ judgments or decisions, independent of whether or not they result in immediate medical errors. Over 32 types of cognitive biases have been described [26]. Importantly, some of these may reflect personality traits that could result in choice tendencies that are factually wrong, whereas others reflect decisions that are potentially suboptimal, although there is no objectively “correct” decision (e.g. risk aversion, tolerance to ambiguity). Both of these factors were included here. Our review has four objectives: 1) to identify the most common cognitive biases by subjecting physicians to real world situations or case-vignettes, 2) to evaluate the influence of cognitive biases on diagnostic accuracy and medical errors in management or treatment, 3) to determine which cognitive biases have the greatest impact on patient outcomes, and 4) to identify literature gaps in this specific area to guide future research. After addressing these objectives, we conclude by highlighting the practical implications of our findings and by outlining an action plan to advance the field.

Methods

Data sources

We conducted a literature search of MEDLINE and the Cochrane Library databases from 1980 to May 2015 by using a pre-specified search protocol (Additional file 1). We used a permuted combination of MeSH terms as major subjects, including: “medical errors”, “bias”, “cognition”, “decision making”, “physicians”, and “case-vignettes” or “case-scenarios”. In-line with the learning and education literature, case-vignettes, clinical scenarios or ‘real world’ encounters are regarded as the best simple strategy to evaluate cognitive biases among physicians [27]. In addition, this approach has also the advantage of facilitating the assessment of training strategies to ameliorate the influence of cognitive biases on medical errors. We therefore restricted our sample to studies that used case-vignettes or real-world encounters. Results of the combination of search terms are listed in the Additional file 1. We also completed further searches based on key words, and reviewed references from previously retrieved articles. All articles were then combined into a single list, and duplicates (n = 106) were excluded (Fig. 2).
Fig. 2

PRISMA flow diagram

PRISMA flow diagram

Study selection

Candidate articles examining cognitive biases influencing medical decisions were included for review if they met the following five inclusion criteria: First, the study was conducted on physicians. Second, at least one outcome measure was reported. Third, at least one cognitive factor or bias was investigated and defined a priori. Fourth, case-vignettes or real clinical encounters were used [28]. Fifth, the study was written in English. We analyzed the number of articles that fulfilled our inclusion criteria on each cognitive factor or bias, methodological aspects, and the magnitude of effect (as prevalence or odds ratios) on diagnostic or therapeutic decisions. We excluded studies that were not the primary source. We analyzed the original data as reported by the authors. Studies not providing raw data were also excluded (e.g. review articles, letters to Editors). A recent systematic review was focused on medical personnel in general rather than physicians, and therefore included a different set of studies in their analysis than those that are of interest when considering the impact of cognitive biases on physicians’ medical decision-making and medical errors (the focus of the current study) [25].

Data extraction

We extracted data according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement (Fig. 2) [29]. Two reviewers (GS, librarian) assessed titles and abstracts to determine eligibility. Data were extracted using standardized collection forms. Information was collected on country of origin, study design, year of publication, number of studied cognitive biases, population target (general practitioners, specialists, residents), decision type (e.g. diagnosis, treatment, management), unadjusted vs. adjusted analysis (for measured confounders, such as age, years of training, expertise), type of outcome (see below), data quality, and summary main findings. We also included descriptive elements (attributes) of the medical information provided for each case-scenario. The main outcomes were any form of medical error [26, 30], including: underuse or overuse of medical tests, diagnostic accuracy, lack of prescription or prescription of unnecessary medications, outcomes of surgical procedures, and avoidable hospitalizations.

Data quality

We used the Newcastle-Ottawa Scale (NOS) to assess the quality of studies (see Additional file 2) [31]. The NOS is a quality assessment tool for observational studies recommended by the Cochrane Collaboration [32]. It assigns one or two points for each of eight items, categorized into three groups: the selection of the study groups; the comparability of the groups; and the ascertainment of the outcome of interest. Previous studies defined NOS scores as: 7–9 points considered as high quality, 5–6 as moderate quality, and 0–4 as low quality [33]. For example, studies that do not provide a description of the cohort, ascertainment of the exposure, adjustment for major confounders, or demonstration that the outcome of interest was not present at the beginning of the study were ranked as low quality [31].

Results

We identified 5963 studies for the combination of MESH terms “decision making” and “physicians”. Of these, 114 fulfilled the selection criteria and were retrieved for detailed assessment. Among them, 38 articles used case-vignettes or real case scenarios in physicians (Fig. 2). Combinations of other search terms are shown in the Additional file 1: Table S1. Twenty studies comprising 6810 physicians (median 180 per study; range: 36–2206) met the inclusion criteria (Fig. 2) [30, 34–52]. In 55 % (n = 11) of the retained studies, results were adjusted for confounders, such as age, gender, level of training (see Additional file 1 for further details). Importantly, only five (25 %) studies used clinical guidelines as the framework to determine diagnostic or treatment errors, illustrating the scarcity of research on evidence-based decision making (e.g. GRADE: decisions based on levels of evidence provided by randomized trials, meta-analysis, etc).

Population target

Eight (40 %) studies included residents, six (30 %) studies included general practitioners, six (30 %) studies included internists, three (15 %) studies included emergency physicians and seven (35 %) studies included other specialists (Table 2). Ten (50 %) studies were conducted in the USA. Only six (30 %) studies classified errors based on real life measures, such as patient encounters, pathological images or endoscopic procedures, whereas the remaining 14 used narrative case-vignettes. Studies included a wide variety of medical situations, most commonly infections (upper respiratory tract, urinary tract) and cardiovascular disease (coronary disease, cerebrovascular disease) (Table 1). In summary, the included studies covered a wide range of medical conditions and participants.
Table 2

Participants, attributes and outcomes of included studies

AuthorType of participantsNumber of vignettes or medical casesNumber of attributesBased on GuidelinesOutcome measureType of outcomea Type of analysisData qualityb Main findings
RedelmeierGPs and Neurologist410-11yesTreatment recommendations4unadjusted5Multiple options decreased the likelihood of medication prescription for pain and carotid endarterectomy by 26 % and 35 %, respectively
RossGPs3NANoDescriptive5adjusted6GPs were less likely to arrange a further consultation for female patients than for male patients (OR = 0.55). GPs with a pessimistic belief about depression were less likely to discuss non-physical symptoms or social factors; More experienced GPs were less likely to conduct a physical examination (OR = 0.60).
GraberGPs28-9NoDescriptive1adjusted4GPs were less likely to believe a serious medical condition among patients with history of depression or somatic symptoms
SorumGPs325yesProbability of ordering a test4adjusted4PSA were more likely ordered among GPs with discomfort for uncertainty and those who expressed regret.
BaldwinPediatric ED physicians397NANoAdmission rates4adjusted5Risk aversion scores higher for physicians with >15 years of experience. Admissions rates did not differ between high and low risk adverse physicians (31.1 vs 30.1; p = 0.91). Adjusted admission rates did not different between high and low discomfort with uncertainty (32.3 vs 29.7; p = 0.84)
FriedmannMedical students (72), residents (72), physicians (72)36 (9)>20NoDiagnostic accuracy5adjusted4Overconfident found in 41 % of residents and in 36 % faculty.
ReynaGPs and specialists9NAYesDiagnostic accuracy and management6adjusted5Physicians deviated from Guidelines in terms of discharge. GP were more risk averse and less likely to discharge patients. Experts achieved better case-risk discrimination by processing less information
BytzerSpecialists5NANoDiagnostic accuracy6unadjusted4Only 23 % endoscopists gave the same diagnosis for the two identical video-cases. The great majority were affected by prior information bias.
DibonaventuraPhysicians211--12NoDescriptive4unadjusted4Naturalness bias present in 40 %, omission bias in 60 % of participants
MamedeResidents8NANo, confirmed diagnosisDiagnostic accuracy5unadjusted5Availability bias increased with years of training. Clinical reasoning ameliorate this bias
Mamadeinternal medicine residents (34) and medical students (50)12>20NoDiagnostic accuracy6unadjusted3Conscious deliberation improved the likelihood of correct diagnosis in physicians, but not in medical students problems were complex, whereas reasoning mode did not matter in simple problems. In contrast, deliberation without attention improved novices’ decisions.
GuptaED Physicians6>20NoDescriptive1adjusted6Outcome bias tends to inflate ratings in the presence of a positive outcome more than it penalizes scenarios with negative ones.
PernegerGPs and specialists, and patients (1121)15NoRating of new drug6adjusted4Physicians and patients provided higher value to the hypothetical new medication when presented in relative terms. Compared to descriptive information, relative mortality reduction (OR 4.40; 3.05 – 6.34), Number needed to treat (OR 1.79; 1.21 – 2.66), and relative survival extension (OR 4.55; 2.74 – 7.55) had a more positive perception.
StieglerResidents (32), Faculty (32)20NACatalogue of common casesManagement1unadjusted41. Developed a cognitive factor/bias catalogue, 2. Top 10 cognitive biases and personality traits: anchoring, availability bias, omission bias, commission bias, premature closure, confirmation bias, framing effect, overconfidence, feedback bias, and sunk cost.3. Errors perceived by faculty to be important to anesthesiology were indeed observed frequently among trainees in a simulated environment.
OgdieResidents41NANoDescriptive6unadjusted3Most common biases: anchoring (88 %), availability (76 %), framing effect (56 %), overconfidence (46 %)
MeyerPhysicians46-11NoDiagnostic accuracy2unadjusted4Higher confidence was related to decreased requests for additional diagnostic tests (P = .01); higher case difficulty was related to more requests for additional reference materials (P = .01).
Crowleypathology residents, fellows and staff pathologists40NANoDiagnostic accuracy6unadjusted4Overall, biases occurred in 52 % of incorrect cases compared to 21 % correct. Most common biases-Availability (20 %) and satisfying biases (22.5 %) the two most common. All the rest, less than 10 %.
SaposnikResidents, internists, emergency physicians and Neurologist105-7NoProbability of death or disability6adjusted5Higher confidence was not associated with better outcome predictions. 70 % of underestimated the risk of the death or disability, 38 % overestimated death at 30 days.
MsaouelResidents24, 5NoDescriptive1adjusted5Gambler’s fallacy in 46 %, conjunction bias 69 %
YeeSpecialists (Obstetricians)3488NANoManagement6adjusted7Physicians with a higher tolerance of ambiguity were less likely to deliver patients by operative vaginal delivery (11.8 % vs 16.4 %; p = 0.006). The effect disappeared in the adjusted analysis (OR 0.77, 95 % CI 0.53-1.1)

NA not available, GP general practitioners

aType of outcome measured: 1 = probability, 2 = rating, 3 = ranking, 4 = yes/no choice, 5 = discrete choice, 6 = more than 2 alternatives

bData quality assessed by the Newcastle-Ottawa Score. See details in the text and Additional file 2

Table 1

Characteristics of studies included in the systematic review

AuthorYear of publicationCountryNumber participantsMethodsClinical problemType of decisionCognitive bias (n)Type of cognitive biasData quality*
Redelmeier1995Canada639SurveyOstoearthritis, TIAManagement and Treatment1Multiple alternative/Decoy bias5
Ross1999UK407SurveyDepressionTreatment and management1Outcome bias6
Graber2000USA232SurveyHeadache, abdominal pain, depressionDiagnosis1Information bias4
Sorum2003USA, France65SurveyProstate cancerDiagnosis1risk aversion4
Baldwin2005USA46ExperimentalBrochiolitisManagement2risk aversion, Ambiguity tolerance5
Friedman2005USA216SurveyNRDiagnosis1Overconfidence4
Reyna2006USA74SurveyUnstable anginaDiagnosis and management1risk aversion5
Bytzer2007Denmark127Video-casesReflux, epigastric painDiagnosis1Infromation bias4
Dibonaventura2008USA2206SurveyImmunizationTreatment2omissions and naturalness bias4
Mamede2010Netherlands36ExperimentHepatitis, IBD, MI, Wernicke, Pneumonia, UTI, MeningitisDiagnosis1Availability, Reflective reasoning5
Mamade2010Netherlands84SurveyAortic dissection, pancreatitis, hepatitis, pericarditis, hyperthiroidism, sarcoidosis, lung cancer, pneumonia, claudication, bacterial endocarditisDiagnosis1Deliveration without attention3
Gupta2011USA587SurveyAbdominal pain, headache, trauma, asthma, chest painDiagnosis1Outcome bias6
Perneger2011Switzerland1439SurveyHIV infectionTreatment-Prognosis1Framing effect4
Stiegler2012USA64Delphi and 38 simulated encountersanaphylaxis, malignant hyperthermia, difficult airway, and pulmonary embolismTreatment and management10anchoring, availability bias, premature closure, feedback bias, framing effect, confirmation bias, omission4
Ogdie2012USA41NarrativesNRDiagnosis9Anchoring, availability, framing effect, blind obedience, confirmation3
Meyer2013USA118SurveyAbdominal pain, headache and rash, fever and arthralgiasDiagnosis1Overconfidence4
Crowley2013International71Pathology casesVesicular and diffuse dermatitidesDiagnosis8anchoring, availability bias, confirmation bias, overconfidence4
Saposnik2013Canada111Case-scenarios from real practiceStrokePrognosis2Overconfidence, anchoring5
Msaouel2014Greece153SurveyTuberculosis, CADDiagnosis2Gambler’s and Conjunction fallacy5
Yee2014USA94ExperimentalDeliveriesManagement and Treatment1Ambiguity tolerance/aversion7

*Data quality assessed using the Newcastle-Ottawa acale (NOS)

Characteristics of studies included in the systematic review *Data quality assessed using the Newcastle-Ottawa acale (NOS)

Data quality

All studies were designed as cohort studies evaluating cognitive biases. According to the NOS, the majority of studies (n = 12, 60 %) were low quality, seven (35 %) studies ranked moderate and only one ranked as high quality [43] (see Additional file 2: Table S2 for details). All studies were classified as representative of the entire population (defined as how likely the exposed cohort was included in the population of physicians).

Presence of most common cognitive biases (Objective 1)

Our first objective was to evaluate the most common cognitive biases affecting physicians’ decisions. Altogether, studies evaluated 19 different cognitive biases (Table 1 and Additional file 1). It is important to bear in mind that these studies do not systematically assess each cognitive bias or personality traits. As a result, it is not possible to provide a true estimate of the prevalence of all cognitive biases among physicians. Overall, at least one cognitive factor or bias was present in all studies. Studies evaluating more than two cognitive biases, found that 50 to 100 % of physicians were affected by at least one [39, 50, 52]. Only three manuscripts evaluated more than 5 cognitive biases in the same study, in-line with the narrow scope of most studies [39, 50, 52]. One third of studies (n = 6) were descriptive, i.e., they provided the frequency of the cognitive bias without outcome data [36, 37, 39, 44, 48, 51]. The most commonly studied personality trait was tolerance to risk or ambiguity (n = 5), whereas the framing effects (n = 5) and overconfidence (n = 5) were the most common cognitive biases. There was a wide variability in the reported prevalence of cognitive biases (Fig. 3). For example, when analyzing the three most comprehensive studies that accounted for several cognitive biases (Fig. 4), the availability bias ranged from 7.8 to 75.6 % and anchoring from 5.9 to 87.8 %, suggestive of substantial heterogeneity among studies. In summary, cognitive biases may be common and present in all included studies. The framing effect, overconfidence, and tolerance to risk/ambiguity were the most commonly studied cognitive biases. However, methodological limitations make it difficult to provide an accurate estimation of the true prevalence.
Fig. 3

Prevalence of most common cognitive biases as reported by different studies. Numbers represent percentages reflecting the frequency of the cognitive factor/bias. Panel a represent the prevalence of the framing effect. Panel b represent the prevalence of prevalence of tolerance to risk and ambiguity. Panel c represents the prevalence of overconfidence. Overall, overconfidence and low tolerance to risk or ambiguity were found in 50-70 % of participants, whereas a wide variation was found for the framing effect

Fig. 4

Prevalence of cognitive biases in the top three most comprehensive studies [39, 50, 52] Numbers represent percentages reflecting the frequency of the cognitive bias. Note the wide variation in the prevalence of cognitive biases across studies

Prevalence of most common cognitive biases as reported by different studies. Numbers represent percentages reflecting the frequency of the cognitive factor/bias. Panel a represent the prevalence of the framing effect. Panel b represent the prevalence of prevalence of tolerance to risk and ambiguity. Panel c represents the prevalence of overconfidence. Overall, overconfidence and low tolerance to risk or ambiguity were found in 50-70 % of participants, whereas a wide variation was found for the framing effect Prevalence of cognitive biases in the top three most comprehensive studies [39, 50, 52] Numbers represent percentages reflecting the frequency of the cognitive bias. Note the wide variation in the prevalence of cognitive biases across studies

Effect of cognitive biases on medical tasks (Objective 2)

Our second objective concerned the assessment of the influence of cognitive biases on diagnostic, medical management or therapeutic tasks. Most studies (12/20; 60 %) targeted cognitive biases in diagnostic tasks, 7 (35 %) studies targeted treatment or management tasks, and 2 studies (10 %) focused on errors in prognosis. The main measure was diagnostic accuracy in 35 % (7/20) of studies (Fig. 5). Overall, the presence of cognitive biases was associated with diagnostic inaccuracies in 36.5 to 77 % of case-scenarios [30, 35, 40, 42, 45, 52, 53]. A study including 71 residents, fellows, and attending pathologists evaluated 2230 skin biopsies with a diagnosis confirmed by a panel of expert pathologists. Information biases, anchoring effects, and the representativeness bias were associated with diagnostic errors in 51 % of 40 case-scenarios (compared to 16.4 % case-scenarios leading to incorrect diagnoses not related to cognitive biases; p = 0.029) [52].
Fig. 5

Outcome measures of studies evaluating cognitive biases. Numbers represent percentages. Total number of studies = 20. Note that 30 % of studies are descriptive and 35 % target diagnostic accuracy. Only few studies evaluated medical management, treatment, hospitalization or prognosis

Outcome measures of studies evaluating cognitive biases. Numbers represent percentages. Total number of studies = 20. Note that 30 % of studies are descriptive and 35 % target diagnostic accuracy. Only few studies evaluated medical management, treatment, hospitalization or prognosis Only seven (35 %) studies provided information to evaluate the association between physicians’ cognitive biases and therapeutic or management errors [38, 41–43, 46, 47, 50]. Five out of the seven (71.4 %) studies showed an association between cognitive biases and these errors [38, 43, 46, 47, 50]. One study showed that overutilization of screening for prostate cancer among healthy individuals was associated with lower aversion to uncertainty (p < 0.01) [46]. In another study including 94 obstetricians who cared for 3488 deliveries, better coping strategies (p < .015) and tolerance to ambiguity (p < .006) were associated with optimal management (reflected by lower instrumental vaginal deliveries) and lower errors [43]. In a study including 32 anesthesiology residents, several cognitive biases (anchoring, overconfidence, premature closure, confirmation bias, etc.) were associated to errors in half of the 38 simulated encounters [50]. Two studies evaluating triage strategies for patients with bronchiolitis and coronary artery disease showed no association between personality traits (e.g. risk aversion or tolerance to uncertainty) and hospital admissions [41, 42]. In summary, our findings suggest that cognitive biases (from one to two thirds of case-scenarios) may be associated with diagnostic inaccuracies. Evidence from five out of seven studies suggests a potential influence of cognitive biases on management or therapeutic errors [38, 43, 46, 47, 50]. Physicians who exhibited information bias, anchoring effects and representativeness bias, were more likely to make diagnostic errors [38, 43, 46, 50]. Further studies are needed to identify what the most common cognitive biases and the most effective strategies to overcome their potential influence of medical tasks and errors.

Effect of physician’s cognitive biases on patient outcomes (Objective 3)

The third objective of the present study was to determine the impact of cognitive biases on patient outcomes (e.g. avoidable hospitalizations, complications related to a procedure or medication, exposure to unnecessary invasive tests, etc). Only two (10 %) studies provided information to answer this question, both evaluating physicians’ tolerance to uncertainty [41, 43]. In a study evaluating obstetrical practices, higher tolerance to ambiguity was associated with an increased risk of postpartum hemorrhage (9.7 % vs 6.5 %; p = .004). The negative effects persisted in the multivariable analysis (for postpartum hemorrhage: OR 1.51, 95 % CI 1.10–2.20 and for chorioamninitis: OR 1.37, 95 % CI 1.10–1.70) [43]. This phenomenon could be explained by overconfidence and underestimation of risk factors associated with maternal infections or puerperal bleeding. On the other hand, a study including 560 infants with bronchiolitis presented to the emergency department cared for by 46 pediatricians showed similar admission rates among physicians with low and high risk aversion or discomfort with diagnostic uncertainty (measured using a standardized tool) [41]. In summary, there too little evidence to make definitive conclusions on the influence of physicians’ personality traits or cognitive biases on patient outcomes.

Literature gaps and recommendations (Objective 4)

We systematically reviewed gaps in the literature. First, most of the studies (60 %) provided a qualitative definition of cognitive biases based on the interpretation of comments made by participants (e.g. illustrative quotes), lacking a unified and objective assessment tool [39, 50]. Second, the unit of study varies from study to study. For example, some authors report results based on the number of physicians involved in the study, whereas others report the results based on the number of case-scenarios. Third, limited information is currently available on the impact of cognitive biases on evidence-based care, as only 15 % of the studies were based on or supported by clinical guidelines (Table 2). Fourth, only one study evaluated the effect of an intervention (e.g. reflective reasoning) to ameliorate cognitive biases in physicians [35]. Fifth, most studies were classified as low quality according to NOS criteria. However, this scale is regarded as having a modest inter-rater reliability. We need consensus among researchers on the best tools to assess the quality of manuscripts. Sixth, only two studies evaluated the influence of physicians’ biases on patient outcomes. Finally, considering the great majority of studies (85 %) targeted only one or two biases (Table 1), the true prevalence of cognitive biases influencing medical decisions remains unknown. Participants, attributes and outcomes of included studies NA not available, GP general practitioners aType of outcome measured: 1 = probability, 2 = rating, 3 = ranking, 4 = yes/no choice, 5 = discrete choice, 6 = more than 2 alternatives bData quality assessed by the Newcastle-Ottawa Score. See details in the text and Additional file 2 As mentioned, medical errors are common in medical practice [5]. Physicians’ biases and personality traits may explain, at least in part, some medical errors. Given the wide practice variability across medical disciplines, decisions on screening tests, surgical procedures, preventative medications, or other interventions (e.g. thrombolysis for acute stroke, antibiotics for an underlying infection, etc.) may not require the same cognitive abilities it is therefore likely that studies from one discipline cannot be transferred automatically to a different discipline. By extension, physicians’ personality traits (e.g. aversion to ambiguity, tolerance to uncertainty) or cognitive biases (e.g. overconfidence) may not equally influence patient outcomes or medical errors in all disciplines. Time-urgency of the medical decision may be a relevant characteristic. Thus, a discipline-based research approach may be needed. There is scarce information in some disciplines and areas, including anesthesiology (decisions on procedures and anesthetic agents), emergency care, obstetrics and gynecology (e.g. decisions on procedures and primary care on women’s health), endoscopic procedures (e.g. gastrointestinal, uropelvic), neurology (e.g. decision in multiple sclerosis and stroke care).

Discussion

Early recognition of physicians’ cognitive and biases are crucial to optimize medical decisions, prevent medical errors, provide more realistic patient expectations, and contribute to decreasing the rising health care costs altogether [3, 8, 54]. In the present systematic review, we had four objectives. First, we identified the most commonly reported cognitive biases (i.e., anchoring and framing effects, information biases) and personality traits (e.g. tolerance to uncertainty, aversion to ambiguity) that may potentially affect physicians’ decisions. All included studies found at least one cognitive factor/bias, indicating that a large number of physicians may be possibly affected [39, 50, 52]. Second, we identified the effect of physician’s cognitive biases or personality traits on medical tasks and on medical errors. Studies evaluating physicians’ overconfidence, the anchoring effect, and information or availability bias may suggest an association with diagnostic inaccuracies [30, 35, 40, 42, 45, 52, 53]. Moreover, anchoring, information bias, overconfidence, premature closure, representativeness and confirmation bias may be associated with therapeutic or management errors [38, 43, 46, 47, 50]. Misinterpretation of recommendations and lower comfort with uncertainty were associated with overutilization of diagnostic tests [46]. Physicians with better coping strategies and tolerance to ambiguity could be related to optimal management [43]. For our third objective – identifying the relation between physicians’ cognitive biases and patient’s outcomes- we only had very sparse data: Only 10 % of studies provided data on this area [41, 43]. Only one study showed higher complications (OR 1.51, 95 % CI 1.10–2.20) among patients cared for by physicians with higher tolerance to ambiguity [43]. The fourth and final objective was to identify gaps in the literature. We found that only few (<50 %) of an established set of cognitive biases [26] were assessed, including: overconfidence, and framing effects. Other listed and relevant biases were not studied (e.g. aggregation bias, feedback sanction, hindsight bias). For example, aggregation bias (the assumption that aggregated data from clinical guidelines do not apply to their patients) or hindsight bias (the tendency to view events as more predictable than they really are) both compromise a realistic clinical appraisal, which may also lead to medical errors [18, 26]. More importantly, only 35 % of studies provided information on the association between cognitive biases or personality traits and medical errors [38, 41–43, 46, 47, 50], with scarce information on their impact on patient outcomes, preventing us from making definite conclusions [41, 43]. Furthermore, the quality of the included studies was classified as low to modest according to NOS criteria, as most studies provided limited descriptions of the exposure and research cohort, and none contributed with follow-up data (e.g. sustainability and reliability of the effects or long-term outcomes) (Additional file 2). When comparing the previous systematic review on patients and medical personnel [25] with ours, some commonalities are apparent. Both reviews agree on the relevance of the topic, identify that a systematic analysis of the impact of cognitive biases on medical decisions is lacking despite substantial work completed in the last two decades [25]. Having a different objective, the authors nicely summarized the number of studies that investigated each cognitive bias either in patients or medical personnel [25]. Similarly, cognitive biases seem to be common among physicians as identified in 80 % (n = 51) of studies included in Blumenthal-Barby and Krieger’s review and all selected studies (n = 20) evaluating at least one outcome in the present review [25]. However, both studies were not able to provide an accurate estimate of the true prevalence of cognitive biases or personality traits affecting medical decisions in physicians. On the other hand, our study adds relevant information regarding the influence of cognitive biases particularly in physicians on diagnostic inaccuracies, suboptimal management and therapeutic errors, and patient outcomes. Our first objective allowed the identification of additional biases (e.g. framing effect, decoy effect, default bias) or physician’s personality traits (e.g. low tolerance to uncertainty, aversion to ambiguity), by including 14 further studies. We also completed a systematic quality assessment of each study using a standardized tool and identified gaps related to the influence of cognitive biases on medical errors [31].

What can be done?

The identification and recognition of literature gaps constitute the first step to finding potential solutions. Increasing awareness among physicians and medical students is an important milestone. A comprehensive narrative review comprising 41 studies on cognitive interventions to reduce misdiagnosis found three main effective strategies: increasing knowledge and expertise, improving clinical reasoning, and getting help from colleagues, experts and tools [55]. First, reflective reasoning counteracts the impact of cognitive biases by improving diagnostic accuracy in second- (OR 2.03; 95 % CI, 1.49–2.57) and first-year residents [OR (odds ratio) 2.31; 95 % CI, 1.89–2.73] [35]. Second, the implementation of tools (e.g. cognitive checklist, calibration) may overcome overconfidence, the anchoring and framing effects (Fig. 5) [8, 9, 56]. Third, heuristics approaches (shortcuts to ignore less relevant information to overcome the complexity of some clinical situations) can improve decision making. As shown by Marewski and Gigerenzer, the identification of three rules (search for predictors to determine their individual importance, stop searching when relevant information was already obtained, and a criteria that specifies how a decision is made) may facilitate prompt decisions and may help physicians to avoid errors in some clinical situations [21, 57, 58]. The inclusion of training in cognitive biases in graduate and postgraduate programs might foster medical education and thereby improve health care delivery [59]. A commitment from academic institutions, scientific organizations, universities, the public, and policy-makers would be needed to reduce a defensive medical practice [60, 61]. An initial step towards this goal may be the ‘Choosing wisely’ strategy [62, 63].

What are the practical implications of our findings?

As shown, cognitive biases and personality traits may affect our clinical reasoning processes which may lead to errors in the diagnosis, management, or treatment of medical conditions [6, 26]. Errors perceived by faculty to be relevant were indeed observed in 50–80 % of trainees in real practice [50]. Misdiagnosis, mismanagement, and mistreatment are frequently associated with poorer outcomes, which are the most common reasons for patients’ dissatisfaction and medical complaints [54, 64, 65]. Our study has several limitations that deserve comment. First, although we aimed to be as systematic as possible in reviewing the literature, we cannot rule out involuntary omissions. It is also possible that our results may be somewhat limited by the strictness of our inclusion criteria. Second, we were not able to complete a formal meta-analysis due to the diversity of definitions and data reported, and small number of studies evaluating specific cognitive biases. In particular, a limited number of studies evaluated the same constructs. Moreover, across studies we often found a lack (in 30 % of studies) or heterogeneity in the outcome measures, mixed denominators (some studies report their findings based on the number of participants, while others based on case-scenarios) [41, 43, 52], and different scope (e.g. some studies are descriptive, [36, 37, 39, 44, 48, 51] whereas others [7, 30, 35, 42, 43, 47, 50, 52] target diagnostic or therapeutic errors). Third, most studies use hypothetical case-vignettes which may not truly reflect medical decisions in real life. Fourth, the assessment of the number of medical elements included in each case scenario may not be consistent (some were reported by authors and others estimated based on the description of case-scenarios) [35, 40, 51]. Fifth, the use of the NOS to assess the quality of studies has been criticized for having modest inter-rater reliability [66, 67]. Despite the aforementioned limitations, our study reflects the relevance and potential burden of the problem, how little we know about the implications of cognitive biases and personality traits on physicians’ decisions, and their impact on patients-oriented outcomes. Our findings may also increase physicians’ awareness of own personality traits or cognitive biases when counseling or advising patients and their family members that may lead to medical errors. From a health policy perspective, this information would provide additional insights on medically relevant cognitive biases and personality traits that contribute the rising health care costs [3, 68].

Conclusions

In the present systematic review, we highlighted the relevance of recognizing physicians’ personality traits and cognitive biases. Although cognitive biases may affect a wide range of physicians (and influence diagnostic accuracy, management, and therapeutic decisions), their true prevalence remains unknown. Thus, substantial gaps limit our understanding of the impact of cognitive biases on medical decisions. As a result, new research approaches are needed. We propose the design of more comprehensive studies to evaluate the effect of physicians’ personality traits and biases on medical errors and patient outcomes in real medical encounters and interventions or using guideline-based case-vignettes. This can be accomplished by identifying physician characteristics, combining validated surveys and experiments commonly used in behavioral economics to elicit several critical personality traits (e.g. tolerance to uncertainty, aversion to risk and ambiguity), and cognitive biases (e.g. overconfidence, illusion of control). Prospective studies evaluating and comparing different training strategies for physicians are needed to better understand and ameliorate the potential impact of cognitive biases on medical decisions or errors. In addition, effective educational strategies are also needed to overcome the effect of cognitive biases on medical decisions and interventions. Together, this information would provide new insights that may affect patient outcomes (e.g. avoidable hospitalizations, complications related to a procedure or medication, request of unnecessary tests, etc) and help attenuate medical errors [3, 68, 69].
  65 in total

1.  Claims, errors, and compensation payments in medical malpractice litigation.

Authors:  David M Studdert; Michelle M Mello; Atul A Gawande; Tejal K Gandhi; Allen Kachalia; Catherine Yoon; Ann Louise Puopolo; Troyen A Brennan
Journal:  N Engl J Med       Date:  2006-05-11       Impact factor: 91.245

2.  Clinical judgment: psychological research and medical practice.

Authors:  A S Elstein
Journal:  Science       Date:  1976-11-12       Impact factor: 47.728

Review 3.  Cognitive diagnostic error in internal medicine.

Authors:  Kees van den Berge; Sílvia Mamede
Journal:  Eur J Intern Med       Date:  2013-04-06       Impact factor: 4.487

Review 4.  Cognitive biases and heuristics in medical decision making: a critical review using a systematic search strategy.

Authors:  J S Blumenthal-Barby; Heather Krieger
Journal:  Med Decis Making       Date:  2014-08-21       Impact factor: 2.583

5.  Analytic methods and medical education. Problems and prospects.

Authors:  A S Elstein
Journal:  Med Decis Making       Date:  1983       Impact factor: 2.583

6.  Conscious thought beats deliberation without attention in diagnostic decision-making: at least when you are an expert.

Authors:  Sílvia Mamede; Henk G Schmidt; Remy M J P Rikers; Eugene J F M Custers; Ted A W Splinter; Jan L C M van Saase
Journal:  Psychol Res       Date:  2010-03-31

7.  Why do primary care physicians in the United States and France order prostate-specific antigen tests for asymptomatic patients?

Authors:  Paul Clay Sorum; Junseop Shim; Gérard Chasseigne; Sylvie Bonnin-Scaon; Joël Cogneau; Etienne Mullet
Journal:  Med Decis Making       Date:  2003 Jul-Aug       Impact factor: 2.583

8.  Information bias in endoscopic assessment.

Authors:  Peter Bytzer
Journal:  Am J Gastroenterol       Date:  2007-08       Impact factor: 10.864

9.  Do decision biases predict bad decisions? Omission bias, naturalness bias, and influenza vaccination.

Authors:  Marco daCosta Dibonaventura; Gretchen B Chapman
Journal:  Med Decis Making       Date:  2008-03-04       Impact factor: 2.583

10.  Measuring the quality of physician practice by using clinical vignettes: a prospective validation study.

Authors:  John W Peabody; Jeff Luck; Peter Glassman; Sharad Jain; Joyce Hansen; Maureen Spell; Martin Lee
Journal:  Ann Intern Med       Date:  2004-11-16       Impact factor: 25.391

View more
  148 in total

1.  Behavioral Heuristics in Coronary-Artery Bypass Graft Surgery.

Authors:  Andrew R Olenski; André Zimerman; Stephen Coussens; Anupam B Jena
Journal:  N Engl J Med       Date:  2020-02-20       Impact factor: 91.245

2.  Combining the Power of Artificial Intelligence with the Richness of Healthcare Claims Data: Opportunities and Challenges.

Authors:  David Thesmar; David Sraer; Lisa Pinheiro; Nick Dadson; Razvan Veliche; Paul Greenberg
Journal:  Pharmacoeconomics       Date:  2019-06       Impact factor: 4.981

3.  Implicit bias in healthcare: clinical practice, research and decision making.

Authors:  Dipesh P Gopal; Ula Chetty; Patrick O'Donnell; Camille Gajria; Jodie Blackadder-Weinstein
Journal:  Future Healthc J       Date:  2021-03

4.  Practicing Medicine with Colleagues: Pitfalls from Social Psychology Science.

Authors:  Donald A Redelmeier; Lee D Ross
Journal:  J Gen Intern Med       Date:  2019-01-31       Impact factor: 5.128

5.  Cognitive Bias in Clinicians' Communication about Human Papillomavirus Vaccination.

Authors:  Caitlin E Hansen; Anna North; Linda M Niccolai
Journal:  Health Commun       Date:  2019-01-24

6.  Comparison of the Informed Health Choices Key Concepts Framework to other frameworks relevant to teaching and learning how to think critically about health claims and choices: a systematic review.

Authors:  Andrew D Oxman; Laura Martínez García
Journal:  F1000Res       Date:  2020-03-05

7.  Hospital-Based Physicians' Intubation Decisions and Associated Mental Models when Managing a Critically and Terminally Ill Older Patient.

Authors:  Shannon Haliko; Julie Downs; Deepika Mohan; Robert Arnold; Amber E Barnato
Journal:  Med Decis Making       Date:  2017-11-22       Impact factor: 2.583

8.  A new approach to integrating patient-generated data with expert knowledge for personalized goal setting: A pilot study.

Authors:  Marissa Burgermaster; Jung H Son; Patricia G Davidson; Arlene M Smaldone; Gilad Kuperman; Daniel J Feller; Katherine Gardner Burt; Matthew E Levine; David J Albers; Chunhua Weng; Lena Mamykina
Journal:  Int J Med Inform       Date:  2020-04-30       Impact factor: 4.046

9.  Predicting seizure freedom after epilepsy surgery, a challenge in clinical practice.

Authors:  Camilo Garcia Gracia; Kevin Chagin; Michael W Kattan; Xinge Ji; Madeleine G Kattan; Lizzie Crotty; Imad Najm; Jorge Gonzalez-Martinez; William Bingaman; Lara Jehi
Journal:  Epilepsy Behav       Date:  2019-04-28       Impact factor: 2.937

10.  User-Centered Clinical Display Design Issues for Inpatient Providers.

Authors:  Thomas A Lasko; David A Owens; Daniel Fabbri; Jonathan P Wanderer; Julian Z Genkins; Laurie L Novak
Journal:  Appl Clin Inform       Date:  2020-10-21       Impact factor: 2.342

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.