| Literature DB >> 33145440 |
Rebecca M Jones1, Anuj Sharma1, Robert Hotchkiss2, John W Sperling3, Jackson Hamburger1, Christian Ledig1, Robert O'Toole4, Michael Gardner5, Srivas Venkatesh1, Matthew M Roberts2, Romain Sauvestre1, Max Shatkhin1, Anant Gupta1, Sumit Chopra1, Manickam Kumaravel6, Aaron Daluiski2, Will Plogger1, Jason Nascone7, Hollis G Potter2, Robert V Lindsey1.
Abstract
Missed fractures are the most common diagnostic error in emergency departments and can lead to treatment delays and long-term disability. Here we show through a multi-site study that a deep-learning system can accurately identify fractures throughout the adult musculoskeletal system. This approach may have the potential to reduce future diagnostic errors in radiograph interpretation.Entities:
Keywords: Bone; Software
Year: 2020 PMID: 33145440 PMCID: PMC7599208 DOI: 10.1038/s41746-020-00352-w
Source DB: PubMed Journal: NPJ Digit Med ISSN: 2398-6352
Fig. 1The deep-learning system.
a The deep-learning system used an ensemble of 10 convolutional neural networks. To produce a prediction, radiographs are processed by each network in the ensemble, averaged, and then post-processed to generate an overall fracture determination and bounding boxes. b Example outputs for each of the 16 anatomical regions supported by the deep-learning system.
Characteristics of the development and test datasets.
| Development dataset | Test dataset | |
|---|---|---|
| Radiographs | ||
| No. of hospitalsa | 15 | 15 |
| No. of radiographs | 715,343 | 16,019 |
| No. of radiographic views | 16b | 9c |
| No. of anatomical regions | 16 | 16 |
| Median (range) radiographs per anatomical region | 40,658 (6249–106,705) | 1000 (774–1079) |
| No. of radiographs with fracture(s) (%) | 82,830 (12%) | 2415 (15%) |
| No. of fracture bounding-box annotations | 97,559 | 2718d |
| No. of bounding-box annotations per fractured radiograph, mean (range) | 1.2 (1–13) | 1.1 (1–6) |
| Patients | ||
| No. of patients | 314,866 | 12,746 |
| Median (range) patients per anatomical region | 18,952 (3022–71,484) | 909 (326–1042) |
| No. of male (%) | 137,929e (44%) | 5520 (43%) |
| Median (range) patient age in years | 54 (0–90)f | 55 (22–90) |
| Annotators | ||
| No. of orthopedic surgeons (median years experience post-residency) | 18 (16) | 11 (13) |
| No. of radiologists (median years experience post-residency) | 11 (13) | 7 (13) |
No radiographs used for testing were in the development dataset.
aDatasets sampled from the MedStar Health System located in Baltimore, MD, Washington, D.C., Olney, MD, Leonardtown, MD, and Clinton, MD, the CarePoint Health System in Bayonne, NJ, Jersey City, NJ, and Hoboken, NJ as well as the Hospital for Special Surgery (HSS) in New York, NY and Orthopedic Institute for Children in CA.
bNumber of unique radiographic views estimated through a manual review of 20,000 randomly sampled radiographs across anatomical regions.
cViews were collapsed for statistical analyses into frontal view (frontal; frontal dorso-plantar; frontal inlet-outlet), lateral view (axillary; frog-leg lateral; lateral; y), and oblique view (oblique; oblique-mortise).
d2718 reflects unique fracture sites after fusing the 3 reference standard annotations per image through a pixel-wise majority vote.
e602 patients were missing biological sex information.
fPatient age missing for 43% of the development dataset because patient age was removed from radiographs collected at HSS. De-identification procedures capped patient age at 90 years. In the development dataset, 0.1% of radiographs were from patients 0 to 10 years of age, and 2.95% were from patients 10 to 20 years of age. By design, no radiographs in the test dataset were from patients <22 years of age.
Fig. 2The deep-learning system’s AUCs.
Error bars represent 95% confidence intervals calculated using bootstrap sampling (m = 1000). n indicates the number of radiographs tested.