| Literature DB >> 34046742 |
Falk Schwendicke1,2, Akhilanand Chaurasia3,4, Lubaina Arsiwala5, Jae-Hong Lee3,6, Karim Elhennawy7, Paul-Georg Jost-Brinkmann7, Flavio Demarco8, Joachim Krois5,3.
Abstract
OBJECTIVES: Deep learning (DL) has been increasingly employed for automated landmark detection, e.g., for cephalometric purposes. We performed a systematic review and meta-analysis to assess the accuracy and underlying evidence for DL for cephalometric landmark detection on 2-D and 3-D radiographs.Entities:
Keywords: Artificial intelligence; Convolutional neural networks; Evidence-based medicine; Meta-analysis; Orthodontics; Systematic review
Mesh:
Year: 2021 PMID: 34046742 PMCID: PMC8310492 DOI: 10.1007/s00784-021-03990-w
Source DB: PubMed Journal: Clin Oral Investig ISSN: 1432-6981 Impact factor: 3.573
Fig. 1Flowchart of the search
Included studies
| 1st author | Year | Country | Imagery | Data source | Architecture/modelling framework | N landmarks | Total sample | Train/validate sample | Reference test on training/validation data | Unification of labels | Test sample | Reference test on test data | Unification of labels |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Arik 2017 [ | 2017 | USA | Lateral 2D | IEEE Grand Challenge 2015 | Custom CNN combined with a shape model for refinement | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
| Chen 2019 [ | 2019 | China | Lateral 2D | IEEE Grand Challenge 2015 | VGG-19, ResNet20, and Inception; custom attentive feature pyramid fusion module | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
| Gilmour 2020 [ | 2020 | Canada | Lateral 2D | IEEE Grand Challenge 2015 | Modified ResNet34 combined with a custom image pyramids approach (spatialized features) | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
| Huang 2020 [ | 2020 | Germany | Lateral 2D | CQ500 CTs (train) and IEEE Grand Challenge 2015 (test) | LeNet-5 for ROI patches and ResNet50 for landmark location | 19 | na | 491 | 3 radiologists | Majority | 150 | 2 experts | Average |
| Hwang 2020 [ | 2020 | Korea | Lateral 2D | Own dataset | Customized YOLO V3 | 80 | 1311 | 1028 | 1 expert | NA | 283 | 1 expert | NA |
| Kim 2020 [ | 2020 | Korea | Lateral 2D | Own dataset+IEEE Grand Challenge 2015 | Stacked hourglass-shaped networks | 23 | 2475 | 1875 | 2 experts | Unclear | 200+225+400 | 2 experts | Unclear or average (IEEE) |
| Lee 2020 [ | 2020 | Korea | Lateral 2D | IEEE Grand Challenge 2015 | Custom CNN for ROI and custom Bayesian CNN for landmark detection | 19 | 400 | 250 | 2 experts | Unclear | 150 | 2 experts | Average |
| Lee 2019 [ | 2019 | Japan | Lateral 2D | Own dataset | Combined custom CNNs for ROI classification and point estimation | 22 | 936 | 835 | 3 experts | Unclear | 100 | 3 experts | Unclear |
| Lee 2019 [ | 2019 | Korea | 3D | Own dataset | VGG-19 | 7 | 27 | 20 | 2 experts | Average | 7 | 2 experts | average |
| Ma [ | 2020 | Japan | 3D | Own dataset | Custom CNNs for classification and regression | 13 | 66 | 58 | 1 expert | NA | 8 | 1 expert | NA |
| Muraev 2020 [ | 2020 | Russia | Frontal 2D | Unclear | Multiclass FPN and ResNeXt-50 with Squeeze-and-Excitation blocks | 45 | 330 | 300 | Students, corrected by experts | Consensus | 30 | students, corrected by experts | Consensus |
| Noothout 2020 [ | 2019 | Netherlands | Lateral 2D | IEEE Grand Challenge 2015 | Custom FCNs based on ResNet34 | 19 | 400 | 150 | 2 experts | Unclear | 150+100 | 2 experts | Average |
| O'Neil 2018 [ | 2018 | UK | Lateral 3D | Own dataset | Custom FCN and Atlas Correction | 22 | 22 | 201 | 3 experts | Unclear | 20 | 2 experts | Unclear |
| Oh 2020 [ | 2020 | Korea | Lateral 2D | IEEE Grand Challenge 2015 | DACFL, custom FCN combined with a local feature perturbator and the anatomical context loss | 19 | 400 | 150 | 2 dental experts | Average | 150+100 | 2 experts | Average |
| Park 2020 [ | 2019 | Korea | Lateral 2D | Own dataset | YOLO V3 and SSD | 80 | 1311 | 1028 | 1 expert | NA | 283 | 1 expert | NA |
| Qian 2020 [ | 2020 | China | Lateral 2D | IEEE Grand Challenge 2015 | Cepha-NN, combining U-Net-shaped networks, attention mechanism, and region enhancing loss | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
| Song 2020[ | 2020 | China | Lateral 2D | IEEE Grand Challenge 2015 | ROI extraction and ResNet50 | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
| Yun 2020 [ | 2020 | Korea | 3D | Own dataset | Custom CNNs, combined skull normalization, and VAE for coarse to fine detection tasks | 93 | 26 | 22 | 1 expert | NA | 4 | 1 expert | NA |
| Zhong 2020 [ | 2019 | China | Lateral 2D | IEEE Grand Challenge 2015 | 2-stage (global and local) U-Net models | 19 | 400 | 150 | 2 experts | Average | 150+100 | 2 experts | Average |
Abbreviations: FCN, fully convolutional neural network; ROI, region of interest. Single Shot Detector.
Risk of bias and applicability concerns
Fig. 2Forest plot of studies reporting the mean deviation from a 2-mm prediction error threshold. Squares indicate the mean deviation of each single study and lines the 95% confidence intervals (95% CI). Yellow and blue diamonds show the pooled subtotal (on 2-D and 3-D imagery) and overall estimates, respectively. I-square and the P value indicate heterogeneity. Studies are ordered according to year; if multiple test datasets were employed in the same study, the second or third is indicated accordingly (e.g., Noothout 2019 (2))
Fig. 3Forest plot of studies reporting the proportion of landmarks correctly predicted within a 2-mm prediction error threshold from the reference. Squares indicate the mean proportion found in each single study and lines the 95% confidence intervals (95% CI). Yellow and blue diamonds show the pooled subtotal (on 2-D and 3-D imagery) and overall estimates, respectively. I-square and the P value indicate heterogeneity. Studies are ordered according to year; if multiple test datasets were employed in the same study, the second or third is indicated accordingly (e.g., Arik 2017 (2))