| Literature DB >> 35933408 |
Rodrigo Echeveste1, Enzo Ferrante2, María Agustina Ricci Lara3,4.
Abstract
Entities:
Mesh:
Year: 2022 PMID: 35933408 PMCID: PMC9357063 DOI: 10.1038/s41467-022-32186-3
Source DB: PubMed Journal: Nat Commun ISSN: 2041-1723 Impact factor: 17.694
Fig. 1Group-fairness metrics.
Here we include a toy-example in the context of disease classification, where two sub-populations characterized by different protected attributes (in red and blue) present different disease prevalence (40% and 20% for blue and red subjects respectively, top row, x marks positive cases). A model optimized for discriminative performance was assessed on a test set achieving 100% accuracy (bottom row left side, + marks positive predictions). Algorithm fairness was audited according to two common metric choices (bottom row, right side). In this case, as a consequence of the difference in disease frequency, the model would not fulfill the demographic parity criterion (bottom row, right side) since the positive prediction rates vary between sub-groups : 40% (8 positive predictions over 20 cases) for the blue sub-group vs. 20% (4 positive predictions over 20 cases) for the red sub-group. On the other hand, the model would fulfill the equal opportunity criterion, as true positive rates match for both sub-groups reaching the value of 100%: 8 true positives out of 8 positive ground truth cases for the blue sub-group and 4 true positives out of 4 positive ground truth cases for the red sub-group . FN false negatives, FP false positives, TN true negatives, TP true positives. See legend-box with symbols on the top right corner.
Fig. 2Main potential sources of bias in AI systems for MIC.
The data being fed to the system during training (1), design choices for the model (2), and the people who develop those systems (3), may all contribute to biases in AI systems for MIC.
Databases commonly used in fairness in MIC studies
| Image modality | Database | Access | Sex or gendera | Age | Skin tone or race/ethnicityb | SES |
|---|---|---|---|---|---|---|
| Chest X-ray | CheXpert[ | Public | x | x | x | – |
| NIH Chest X-Ray[ | Public | x | x | – | – | |
| MIMIC Chest X-Ray[ | Public | x | x | x | x | |
| Emory University Hospital Chest X-Ray[ | Private | x | x | x | – | |
| Mammography | Digital Mammographic Imaging Screening Trial (DMIST)[ | Private | x | x | x | – |
| Emory University Hospital Mammography[ | Private | x | x | x | – | |
| Dermoscopy | ISIC Challenge 2017/18/20[ | Public | x | x | – | – |
| Dermatological clinical image | Fitzpatrick 17k[ | Public | – | – | x | – |
| SD-198[ | Public | – | – | – | – | |
| Fundus image | AREDS[ | Public | x | x | x | – |
| Kaggle EyePACS[ | Public | – | – | – | – | |
| Cardiac MRI | UK Biobank[ | Public | x | x | x | x |
| Pulmonary angiography CT | Stanford University Medical Center[ | Public | x | x | x | – |
aAccording to the World Health Organization, sex refers to different biological and physiological characteristics of males and females, while gender refers to the socially constructed characteristics of women and men such as norms, roles and relationships of and between groups of women and men. Databases tend to report one or the other.
bWe include both the term race and ethnicity since the cited studies make use of both denominations. We group analyses across different skin tones in this category as well. Race and ethnicity are social constructs with complex and dynamic definitions (see ref. 47).