Literature DB >> 36209283

Deep learning model for analyzing the relationship between mandibular third molar and inferior alveolar nerve in panoramic radiography.

Shintaro Sukegawa1,2,3, Futa Tanaka4, Takeshi Hara4,5, Kazumasa Yoshii4, Katsusuke Yamashita6, Keisuke Nakano7, Kiyofumi Takabatake7, Hotaka Kawai7, Hitoshi Nagatsuka7, Yoshihiko Furuki8.   

Abstract

In this study, the accuracy of the positional relationship of the contact between the inferior alveolar canal and mandibular third molar was evaluated using deep learning. In contact analysis, we investigated the diagnostic performance of the presence or absence of contact between the mandibular third molar and inferior alveolar canal. We also evaluated the diagnostic performance of bone continuity diagnosed based on computed tomography as a continuity analysis. A dataset of 1279 images of mandibular third molars from digital radiographs taken at the Department of Oral and Maxillofacial Surgery at a general hospital (2014-2021) was used for the validation. The deep learning models were ResNet50 and ResNet50v2, with stochastic gradient descent and sharpness-aware minimization (SAM) as optimizers. The performance metrics were accuracy, precision, recall, specificity, F1 score, and area under the receiver operating characteristic curve (AUC). The results indicated that ResNet50v2 using SAM performed excellently in the contact and continuity analyses. The accuracy and AUC were 0.860 and 0.890 for the contact analyses and 0.766 and 0.843 for the continuity analyses. In the contact analysis, SAM and the deep learning model performed effectively. However, in the continuity analysis, none of the deep learning models demonstrated significant classification performance.
© 2022. The Author(s).

Entities:  

Mesh:

Year:  2022        PMID: 36209283      PMCID: PMC9547920          DOI: 10.1038/s41598-022-21408-9

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

Third molar extraction is the most common surgery performed by dentists and maxillofacial surgeons. The mandibular third molar has more complications than the maxillary third molar, including post-extraction infection, postoperative pain, and inferior alveolar nerve damage[1,2]. Among these complications, inferior alveolar nerve damage should be avoided because it stresses patients for a prolonged period. In clinical practice, panoramic radiographs are generally used to determine the difficulty of the mandibular third molar, including the contact with the inferior alveolar nerve, depth of the mandibular third molar, and distance to the mandibular ramus. If contact between the inferior alveolar canal and mandibular third molar is suspected after the panoramic screening, computed tomography (CT) is used to identify defects in the cortical bone around the inferior alveolar nerve. The cortical bone defects are significantly risky for postoperative inferior alveolar nerve damage[3]. Accurately determining the positional relationship between the inferior alveolar nerve and the mandibular third molar with a two-dimensional panoramic radiograph is difficult, but preoperative diagnosis using CT in 3D is very effective[4]. Although CT cannot directly image the nerve, it can clarify the positional relationship between the tooth and the nerve by depicting the inferior alveolar nerve and the bony border[5]. However, CT imaging cannot be applied in all cases owing to radiation exposure and high cost[6]. Therefore, developing an assistant diagnostic tool to diagnose the contact relationship with the inferior alveolar nerve from panoramic radiograph images is essential. Convolutional neural networks (CNN) have revolutionized deep learning in recent years. CNN-based classifiers have proved highly accurate for image recognition[7] and have consequently impacted diagnostic imaging in the medical field. They have been applied to the detection of lung cancer from chest X-ray images[8], determination of retinal detachment[9], detection of osteoporosis[10], screening of breast cancer[11], etc. In addition, many deep learning-related studies have been reported in the field of dentistry, and classifiers have been developed for areas such as caries[12], periapical lesions[13], dental implants[14], maxillary sinusitis[15], and position classification of the mandibular third molars[16]. Furthermore, deep learning has occasionally been more accurate than human diagnosis[17,18]. In contact analysis using deep learning, Fukuda et al.[19] examined images of different sizes and reported that the results were more accurate when the images were small and condensed only to those that required a large amount of information. Thus, deep learning diagnostic imaging has great potential and must be explored. We therefore hypothesized that deep learning could accurately diagnose the positional relationship between the mandibular third molar and inferior alveolar nerve on panoramic radiographs. This study aimed to explain the accuracy of the positional relationship of the contact between the inferior alveolar canal and the mandibular third molar using deep learning. To this end, in this CNN deep learning-based study, we first investigated the diagnostic performance of the presence or absence of contact between the mandibular third molar and inferior alveolar canal. Subsequently, we explored the diagnostic performance of bone continuity between the mandibular third molar and inferior alveolar canal, diagnosed using CT.

Materials and methods

Study design

This study analyzed the diagnostic performance of the positional relationship between the inferior alveolar canal/nerve and the mandibular third molar from panoramic radiographs using an optimized CNN deep learning model.

Ethics statement

This study was approved by the Institutional Review Board of Kagawa Prefectural Central Hospital (approval number: 1023; approval date: 8th March 2021). The board reviewed our retrospective non-interventional study design and analytical study with anonymized data and waived written documentation of personal informed consent. All methods were performed following the relevant guidelines and regulations. The study was registered at jRCT (jRCT1060220021).

Preparation of image datasets

We retrospectively used radiographic imaging data collected at the Department of Oral and Maxillofacial Surgery in a single general hospital from April 2014 to December 2021. The study data included patients aged 20–76 years in the mature mandibular third molar who had panoramic radiographs and CT taken on the same day. This study confirms the positional relationship between the mandibular third molar and the inferior alveolar canals by panoramic radiography. An unclear image (three teeth) and an image of the remaining titanium plate after the mandibular fracture (one tooth) were excluded. Finally, 1279 tooth images were used in this study. Digital image data were obtained using dental panoramic radiographs taken with either of the two imaging devices (AZ3000CMR or Hyper-G CMF; ASAHIRENTOGEN Ind. Co., Ltd., Kyoto, Japan). All digital image data were output in a tagged image file format (digital image size: 2776 × 1450, 2804 × 1450, 2694 × 1450, or 2964 × 1464 pixels) using the Kagawa Prefectural Central Hospital Picture Archiving and Communication Systems system (Hope Dr. Able-GX, Fujitsu Co., Tokyo, Japan). Under the supervision of an expert oral and maxillofacial surgeon, two oral and maxillofacial surgeons used Photoshop Elements (Adobe Systems, Inc., San Jose, CA, USA) to crop the areas of interest manually. The image was cropped by selecting the area, including the apex of the mandibular third molar and the inferior alveolar canal within 250 × 200 pixels (Fig. 1). Each cropped image had a resolution of 96 dpi and was saved in the portable network graphic format.
Figure 1

Classification of the relationship between the mandibular third molar and inferior alveolar canal/nerve.

Classification of the relationship between the mandibular third molar and inferior alveolar canal/nerve.

Classification of mandibular third molar and inferior alveolar nerve

First, using panoramic radiographs, we classified the contact and superimposition between the mandibular third molar and inferior alveolar canal. This is because contact between the inferior alveolar duct on panoramic radiographs is a risk factor for nerve exposure[4,20]. Second, we classified the presence or absence of direct contact between the mandibular third molar and inferior alveolar nerve using CT. The classification criteria and distributions are as follows and are also shown in Fig. 1. Relationship between the mandibular third molar and inferior alveolar canal. Non-contact or superimposition of the mandibular third molar and inferior alveolar canal. Contact or superimposition between the mandibular third molar and the inferior alveolar canal. In this study, contacts and superimposition/overlaps were grouped together. Relationship between the mandibular third molar and inferior alveolar nerve. If there was discontinuity of the cortical bone at the inferior alveolar canal due to the mandibular third molar, it was classified as a defect. Contact between the mandibular third molar and inferior alveolar nerve (i.e., defect or discontinuity in the cortical bone of the inferior alveolar canal). Non-contact between the mandibular third molar and inferior alveolar nerve (i.e., continuity of the cortical bone of the inferior alveolar canal).

CNN model architecture

ResNet50 is a 50-layer deep CNN model. Traditional CNNs have the major drawback of the “vanishing gradient problem,” where the gradient value is significantly reduced during backpropagation, resulting in little weight change. The ResNet CNN model uses a residual module to overcome this problem[21]. ResNet v2 is an improved version of the original ResNet[22], with the following improvements compared with the original ResNet (Fig. 2): (1) The shortcut path is completely identity mapped without using the ReLU between the input and output. (2) After branching for the residual calculation, the order is changed to batch normalization[23] as -ReLU-convolution-batch normalization-convolution.
Figure 2

Differences between the residual blocks of ResNet and ResNetv2: (a) ResNet Residual Unit; (b) ResNetv2 Residual Unit. BN: Batch Normalization and Conv2D: Two-dimensional convolution layer.

Differences between the residual blocks of ResNet and ResNetv2: (a) ResNet Residual Unit; (b) ResNetv2 Residual Unit. BN: Batch Normalization and Conv2D: Two-dimensional convolution layer. In this study, we selected two CNN models, ResNet50 and ResNet50v2. The ResNet50 and ResNet50v2 CNN models were pre-trained on the ImageNet database and fine-tuned according to the positional relationship classification task for the mandibular third molar and inferior alveolar nerve. The deep learning task process was implemented using Python (version 3.7.13), Keras (version 2.8.0), and TensorFlow (version 2.8.0).

Dataset and CNN model training

Each CNN model training was generalized using K-fold cross-validation in the deep learning algorithm. The models were validated using tenfold cross-validation to ensure internal validity. The digital image dataset was divided into ten random subsets using the stratified sampling technique, and the same classification distribution was maintained for training, validation, and testing across all subsets[24]. The dataset was split into separate test and training datasets in a ratio of 0.1–0.9 within each fold. Additionally, the validation data comprised one-tenth of the training dataset. The model averaged ten training iterations to obtain prediction results for the entire dataset, with each iteration retaining a different subset for validation. The cross-entropy—defined by Eq. (1)—was used for the loss function.where t is true label and y is the predicted probability of class i.

Optimization algorithm

This study used two deep learning gradient methods, stochastic gradient descent (SGD) and sharpness-aware minimization (SAM). SGD is a typical optimization method in which the parameters are updated by the magnitude in the obtained gradient direction. The momentum SDG is a method of adding momentum to SGD[25]. In this study, the momentum was set to 0.9. The momentum SGD is expressed in Eqs. (2) and (3).where w_t is t-th parameter, η is learning rate, ∇L (w) is differentiation with parameters of the loss function, and α is momentum. SAM is an optimization method that converges to a parameter with minimal loss and flat surroundings[26]. It uses a combination of a base optimizer and SAM to determine the final parameters using traditional algorithms. SGD was selected as the base optimizer. The loss function of SAM is defined by Eq. (4). This is used to minimize Eq. (5).where S is the set of data, w is a parameter, λ is the L2 regularization coefficient, L_s is the loss function, and ρ is the neighborhood size. This study analyzed the deep learning models using a ρ value of 0.025.

Deep learning procedure

Data augmentation

Data augmentation prevents excessive adaptation to the training data by diversifying the input data[27]. The following values were selected for the preprocessing layer to convert the images during training randomly. The boundary surface of the missing part was complemented by folding back using the reflect method. Random rotation: range of − 18° to 18° Random flip: horizontally and vertically Random translation: up–down and left–right range of 30%

Learning rate scheduler

Learning rate decay is a technique used to improve the generalization performance of deep learning and reduce the learning rate from a state in which learning has progressed to some extent. Decay in the learning rate can improve accuracy[21]. The changes due to time-based decay as a learning rate can be found in the appendix. The learning rate decay can be evaluated using Eq. (6).The learning rate scheduler was executed with an initial learning rate of 0.01 and a decay rate of 0.001. All the models conducted analysis over 300 epochs and with 32 batch sizes without early stopping. These deep learning processes were repeated 30 times for all models using different random seeds for each analysis.

Performance metrics and statistical analysis

To evaluate the performance of each deep learning model, the accuracy, precision, recall, F1 score, and area under the curve (AUC)—calculated from the receiver operating characteristic (ROC) curve—performance metrics were employed. More detailed information on the performance metrics used in this study is present in the Appendix. Statistical evaluations of the performance for each deep learning model were performed on the data that were independently and repeatedly analyzed 30 times. Data were recorded and stored in an electronic database using Microsoft Excel (Microsoft Inc., Redmond, WA, USA). The database was created and analyzed by using JMP Statistical Software Package Version 14.2.0 for Macintosh (SAS Institute Inc., Cary, NC, USA). All statistical analyses were bilateral with a significance level of 0.05. Normal distribution was evaluated by using the Shapiro–Wilk test. A comparison of classification performance between each CNN model was performed for each metric by using the Wilcoxon signed rank sum test. Effect sizes[28] were evaluated using Hedges' g (unbiased Cohen's d), Eqs. (7) and (8).where M1 and M2 are the mean values for the CNN models with SGD and SAM, s1 and s2 are the standard deviations for the CNN models with SGD and SAM, respectively; and n1 and n2 are the numbers for the CNN models with SGD and SAM, respectively. Effect sizes were categorized as large effect, ≥ 2.0; very large effect, 1.0; large effect, 0.8; medium effect, 0.5; small effect, 0.2; and very small effect, 0.01 based on the criteria proposed by Cohen and extended by Sawilowsky[29].

Visualization of judgment regions in deep learning

In this study, the gradient-weighted class activation map (Grad-CAM) algorithm[30] was used to visualize the noticeable areas of the image in a heatmap. Grad-CAM is a class activation mapping method that uses gradients for weights adopted by the IEEE International Conference on Computer Vision in 2017 and provides a visual basis for deep learning to improve the explanation of the architecture. Grad-CAM uses the last convolution layer of the ResNet model to visualize the feature area.

Results

Performance metrics of ResNet50 and ResNet50v2 in the SAM and SGD optimizers

Table 1 shows the results of the performance metrics of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in the contact analysis. In the contact analysis of the inferior alveolar canal and mandibular third molar on panoramic radiographic images, ResNet 50v2 using the SAM optimizer showed the highest performance on all performance metrics (Accuracy: 0.860, Precision: 0.816, Recall: 0.791, F1 score: 0.800, and AUC: 0.890).
Table 1

Performance metrics of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in contact analysis.

CNNOptimizerAccuracyPrecisionRecallF1 scoreAUC
SDSDSDSDSD
95%CI95% CI95% CI95% CI95% CI
ResNet50SAM0.8550.8100.7850.7940.883
0.0050.0090.0090.0080.007
0.853–0.8570.807–0.8130.782–0.7890.791–0.7970.880–0.885
ResNet50SGD0.8500.8040.7810.7890.875
0.0090.0100.0100.0090.008
0.847–0.8530.800–0.8070.785–0.7780.786–0.7930.872–0.878
ResNet50v2SAM0.8600.8160.7910.8000.890
0.0050.0080.0090.0080.007
0.858–0.8610.813–0.8190.788–0.7940.798–0.8030.888–0.893
ResNet50v2SGD0.8530.8090.7820.7920.884
0.0050.0080.0090.0070.006
0.851–0.8550.806–0.8120.779–0.7850.790–0.7950.882–0.886

SD, standard deviation; 95% CI, 95% confidence interval; AUC, area under the receiver operating characteristics curve.

Performance metrics of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in contact analysis. SD, standard deviation; 95% CI, 95% confidence interval; AUC, area under the receiver operating characteristics curve. Table 2 shows the results of the performance metrics of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in the continuity analysis. In the continuity analysis of the inferior alveolar nerve and mandibular third molar on panoramic radiographic images, ResNet50v2 using the SAM optimizer showed the highest performance on all performance metrics and contact analysis (Accuracy: 0.766, Precision: 0.766, Recall: 0.765, F1 score: 0.775, and AUC: 0.843).
Table 2

Performance metrics of ResNet50 and ResNet50v2 with optimizers SAM and SGD continuity analysis.

CNNOptimizerAccuracyPrecisionRecallF1 scoreAUC
SDSDSDSDSD
95% CI95% CI95% CI95% CI95% CI
ResNet50SAM0.7540.7550.7540.7530.832
0.0050.0080.0080.0080.006
0.753–0.7560.752–0.7570.751–0.7570.750–0.7550.829–0.834
ResNet50SGD0.7540.7540.7540.7520.830
0.0070.0080.0080.0080.006
0.752–0.7570.752–0.7570.751–0.7570.750–0.7550.827–0.832
ResNet50v2SAM0.7660.7660.7650.7750.843
0.0070.0060.0060.0130.005
0.764–0.7690.764–0.7680.763–0.7670.771–0.7800.842–0.845
ResNet50v2SGD0.7650.7650.7650.7670.842
0.0060.0060.0060.0130.005
0.763–0.7680.763–0.7670.762–0.7670.762–0.7720.840–0.844

SD, standard deviation; 95% CI, 95% confidence interval; AUC, area under the receiver operating characteristics curve.

Performance metrics of ResNet50 and ResNet50v2 with optimizers SAM and SGD continuity analysis. SD, standard deviation; 95% CI, 95% confidence interval; AUC, area under the receiver operating characteristics curve.

Statistical evaluation of performance metrics in each CNN model

Tables 3 and 4 show the statistical evaluation results of both CNN models for each performance metric. Contact and continuity analyses yielded symmetrical results. For the contact analysis results shown in Table 3, both ResNet50 and ResNet50v2 exhibited statistically significant differences on all performance metrics for SAM and SGD. AUC and accuracy for ResNet50 showed the highest effect size equivalent to “very large” using SAM. The comparison of ResNet50v2 and ResNet50 using SAM showed a statistically significantly higher performance for ResNet50v2 on all performance metrics.
Table 3

Statistical evaluation of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in contact analysis.

Performance metricsModel AModel BA-BP valueEffect size
ResNet50
AccuracySAMSGD0.0060.0030.761
Precision0.0060.0060.677
Recall0.0040.0460.440
F1 score0.0050.0180.556
AUC0.008 < .00011.052
ResNet50v2
AccuracySAMSGD0.007 < .00011.456
Precision0.007 < .00010.874
Recall0.009 < .00010.995
F1 score0.008 < .00011.103
AUC0.0060.0010.899
ResNet50v2 versus ResNet50 optimizer; SAM
AccuracyResNet50v2ResNet500.0040.0040.835
Precision0.0060.0040.742
Recall0.0050.0130.560
F1 score0.0060.0040.712
AUC0.007 < .00010.932

AUC, area under the receiver operating characteristics curve.

Table 4

Statistical evaluation of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in continuity analysis.

Performance metricsModel AModel BA-BP valueEffect size
ResNet50
AccuracySAMSGD0.00010.84420.0162
Precision0.00010.86290.0109
Recall0.00010.84850.0119
F1 score0.00020.81760.0195
AUC0.00220.09960.3512
ResNet50v2
AccuracySAMSGD0.00070.63280.1103
Precision0.00100.53890.1576
Recall0.00010.91230.0214
F1 score0.00800.05230.6064
AUC0.00140.25840.2737
ResNet50 versus ResNet50v2 optimizer; SAM
AccuracyResNet50v2ResNet500.0116 < .00011.9765
Precision0.0112 < .00012.3677
Recall0.0105 < .00012.0622
F1 score0.0225 < .00014.6346
AUC0.0113 < .00012.1598

AUC, area under the receiver operating characteristic curve.

Statistical evaluation of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in contact analysis. AUC, area under the receiver operating characteristics curve. Statistical evaluation of ResNet50 and ResNet50v2 with the SAM and SGD optimizers in continuity analysis. AUC, area under the receiver operating characteristic curve. For the continuity analysis shown in Table 4, neither ResNet50 nor ResNet50v2 demonstrated a statistically significant difference on any performance metric when comparing SAM and SGD. The effect size was “small” to “very small” for ResNet50. Conversely, a comparison of ResNet50v2 and ResNet50 using SAM demonstrated a statistically higher performance for ResNet50v2 on all performance metrics, and all effect sizes also showed “very large” to “huge.”

Comparison of the learning curves of the CNN models

Figure 3 shows the learning curve for each CNN deep learning model. In the contact analysis, SGD exhibited a tendency for overfitting with increasing epochs, whereas for the CNN model with SAM, SAM exhibited low overfitting. Interestingly, continuity analysis also demonstrated overfitting for the CNN models using SAM.
Figure 3

Learning curves for each CNN model in contact and continuity analyses.

Learning curves for each CNN model in contact and continuity analyses.

Visualization of model classification by Grad-CAM

Figure 4 shows the visualization of the area of interest for classification decisions in each deep learning model in the contact and continuity analyses. In the ResNet50 and ResNet50v2-based CNN models, Grad-CAM visualized the final layer of the convolutional layer or the feature area using a heat map. There was no significant difference in the feature areas indicated by the Grad-CAM in contact and continuity analyses. The point of contact between the inferior alveolar canal and mandibular third molar or the closest part was determined to be the characteristic area. This area of interest was the same as the dentist's judgment. In the heatmap visualization using Grad-CAM, the warmer the color, the more significant the contribution to feature determination.
Figure 4

Visualization of regions of interest for CNN classification in contact and continuity analyses.

Visualization of regions of interest for CNN classification in contact and continuity analyses.

Discussion

In this study, we analyzed deep learning models using contact and continuity analyses to classify the positional relationship of the inferior alveolar canal. The CNN model demonstrated a high performance in contact analysis. The CNN model using SAM as the optimizer for ResNet 50v2 exhibited the highest performance. However, in the continuity analysis, none of the CNN models showed high classification performance. This study focused on the part in contact with the inferior alveolar canal and mandibular third molar and obtained a high classification performance using CNN models. However, some cases have been misclassified due to the tooth-like sclerotic area between the teeth and the inferior alveolar canal, and the cropped images made it difficult to identify the inferior alveolar canal. Therefore, optimized data collection is required for contact analysis. In general, as CT and MRI can provide three-dimensional (3D) information, it is possible to accurately determine the positional relationship between the lower alveolar canal and the mandibular third molar by comparing it with a panoramic image[31]. Analysis using deep learning was performed to determine the 3D positional relationships from panoramic images without using other imaging devices. In continuity analysis, although it is difficult to simply compare performance in deep learning studies conducted on different data, a deep learning classification study conducted on 571 images by Choi et al. reported a classification accuracy of 0.723[32]. The accuracy of the deep learning model in this study was 0.766, which is almost the same as the classification accuracy. In addition, the diagnostic accuracy of specialists was 0.55–0.72 (average 0.63), and it was difficult for even specialists to evaluate the continuity between the inferior alveolar nerve and the mandibular third molar using only panoramic radiographic images. The diagnostic accuracy of deep learning is also equivalent to the highest value among specialists, suggesting CNN models cannot improve the breakthrough diagnosis of continuity. SGD identifies a point that minimizes the loss function. Although the loss function becomes small, the peripheral optimization parameters become nonuniform. This leads to overfitting and reduced generalization performance. In contrast, in SAM, the loss function is designed to search for flat parameters. Therefore, the values around the selected parameters also exhibited a uniformly low loss function. It improves the generalization performance and robustness against noise. In this study, SGD showed a tendency for overfitting in the contact analysis in comparison to SAM. By contrast, the continuity analysis showed a trend to overfit even in the CNN model using SAM. This is probably due to inconsistencies between the image data and the correct label. In other words, it suggests that even with the deep learning method, it was not possible to identify the absolute feature showing continuity of the inferior alveolar canals. This is the first study to analyze the relationship between the inferior alveolar nerve and mandibular third molars using SAM. The findings of this study will contribute to the development of deep learning in dentistry in the future. The characteristics of ResNet, a derivative of ResNet50, are (1) input batch normalization and ReLU activation before the convolution operation and (2) nonlinearity creation as an identity mapping. In other words, the output of the additive operation between the identity mapping and residual map can be passed directly to the next block for further processing to facilitate the propagation of information. In this study, the learning curve of ResNet50v2 exhibited a more stable learning process than that of ResNet50. In addition, ResNet50v2 showed a statistically significant improvement in performance metrics in both the contact and continuity analyses, demonstrating that it is an optimal CNN model. One problem with deep learning is that the inference process for the input data is a black box and the reason for extracting the features cannot be explained. Model output explanation has been proposed as an approach for explainable artificial intelligence[33], to explain the rationale for predicting the output of deep learning. Grad-CAM and guided Grad-CAM are class activation mapping methods that use gradients and are often used in deep learning of medical images[34,35]. In the Grad-CAM used in this study, the focus area was the contact between the inferior alveolar canal and the mandibular third molar or the closest site in both the contact and continuity analyses. In other words, it is likely that learning is possible with an understanding of the exact feature area. However, it is difficult to understand the characteristic areas at more detailed points, such as the defect of the cortical bone on the upper wall of the inferior alveolar canal. Thus, further research is required to examine the approach of explainable AI with guided backprop[36]. This study has several strengths. First, we analyzed the positional relationship between the inferior alveolar canal and the mandibular third molar using panoramic radiographs. We evaluated the continuity of the inferior alveolar canal with contact and CT findings as correct labels from the panoramic radiograph findings. This study was the first to use the same image to determine the ability to classify positional relationships using panoramic radiographs as screening and determine whether to classify the position of the inferior alveolar nerve in three dimensions as a potential evaluation of deep learning. Second, this study is the first to introduce effect size as a statistical evaluation method for the performance metrics of each CNN model that analyzes the positional relationship between the inferior alveolar canal and mandibular third molar. The effect size indicates effectiveness of an analytical operation and the strength of the association between each variable[37]. Therefore, the detected effect size is an essential prior parameter to help estimate the sample size in studying the relationship between the inferior alveolar canal and mandibular third molar using deep learning. This study had two limitations. The first is that the data were collected from a single facility and were not validated externally. Internal validity was evaluated using confidence intervals from the dataset via mutual validations. However, to satisfy the external validity criterion, more data must be used in multicenter joint research. The second is the use of only two CNN models. In this study, we analyzed the data using ResNet50 and ResNet50v2. By examining previously published CNN and original CNN models, it may be possible to identify a better model for classifying the relationship between the inferior alveolar canal and the mandibular third molar.

Conclusions

In this study, we investigated the effects of a deep learning model using contact and continuity analyses to classify the positional relationship of the inferior alveolar canal. Contact analysis classified the presence or absence of contact with the mandibular third molar on panoramic radiographs, and continuity analysis classified the presence or absence of bone continuity in the inferior alveolar canal on CT images. CNN models showed a high performance in contact analysis. The CNN deep learning model using SAM as the optimizer for ResNet50v2 exhibited the highest performance. However, the continuity analysis did not show high classification performance. These results indicate that deep learning plays a vital role in primary screening using panoramic radiographs to evaluate the positional relationship between the inferior alveolar canal and the mandibular third molar. However, further studies are needed on deep learning to replace CT imaging for 3D evaluation. Supplementary Information.
  25 in total

1.  A deep transfer learning approach for the detection and diagnosis of maxillary sinusitis on panoramic radiographs.

Authors:  Mizuho Mori; Yoshiko Ariji; Akitoshi Katsumata; Taisuke Kawai; Kazuyuki Araki; Kaoru Kobayashi; Eiichiro Ariji
Journal:  Odontology       Date:  2021-05-23       Impact factor: 2.634

2.  Mandibular canal-related parameters interpreted in panoramic images and CBCT of mandibular third molars as risk factors to predict sensory disturbances of the inferior alveolar nerve.

Authors:  L H Matzen; L B Petersen; L Schropp; A Wenzel
Journal:  Int J Oral Maxillofac Surg       Date:  2019-04-08       Impact factor: 2.789

3.  Comparison of 3 deep learning neural networks for classifying the relationship between the mandibular third molar and the mandibular canal on panoramic radiographs.

Authors:  Motoki Fukuda; Yoshiko Ariji; Yoshitaka Kise; Michihito Nozawa; Chiaki Kuwada; Takuma Funakoshi; Chisako Muramatsu; Hiroshi Fujita; Akitoshi Katsumata; Eiichiro Ariji
Journal:  Oral Surg Oral Med Oral Pathol Oral Radiol       Date:  2020-05-19

4.  Deep Learning to Distinguish Recalled but Benign Mammography Images in Breast Cancer Screening.

Authors:  Sarah S Aboutalib; Aly A Mohamed; Wendie A Berg; Margarita L Zuley; Jules H Sumkin; Shandong Wu
Journal:  Clin Cancer Res       Date:  2018-10-11       Impact factor: 12.531

Review 5.  Artificial Intelligence Versus Clinicians in Disease Diagnosis: Systematic Review.

Authors:  Jiayi Shen; Casper J P Zhang; Bangsheng Jiang; Jiebin Chen; Jian Song; Zherui Liu; Zonglin He; Sum Yi Wong; Po-Han Fang; Wai-Kit Ming
Journal:  JMIR Med Inform       Date:  2019-08-16

6.  Efficacy of proteolytic enzyme bromelain on health outcomes after third molar surgery. Systematic review and meta-analysis of randomized clinical trials.

Authors:  M-L-T Mendes; E-M do Nascimento-Júnior; D-M Reinheimer; P-R-S Martins-Filho
Journal:  Med Oral Patol Oral Cir Bucal       Date:  2019-01-01

7.  Evaluation of multi-task learning in deep learning-based positioning classification of mandibular third molars.

Authors:  Shintaro Sukegawa; Tamamo Matsuyama; Futa Tanaka; Takeshi Hara; Kazumasa Yoshii; Katsusuke Yamashita; Keisuke Nakano; Kiyofumi Takabatake; Hotaka Kawai; Hitoshi Nagatsuka; Yoshihiko Furuki
Journal:  Sci Rep       Date:  2022-01-13       Impact factor: 4.379

8.  The Clinical Value of Explainable Deep Learning for Diagnosing Fungal Keratitis Using in vivo Confocal Microscopy Images.

Authors:  Fan Xu; Li Jiang; Wenjing He; Guangyi Huang; Yiyi Hong; Fen Tang; Jian Lv; Yunru Lin; Yikun Qin; Rushi Lan; Xipeng Pan; Siming Zeng; Min Li; Qi Chen; Ningning Tang
Journal:  Front Med (Lausanne)       Date:  2021-12-14

9.  Identification of osteoporosis using ensemble deep learning model with panoramic radiographs and clinical covariates.

Authors:  Shintaro Sukegawa; Ai Fujimura; Akira Taguchi; Norio Yamamoto; Akira Kitamura; Ryosuke Goto; Keisuke Nakano; Kiyofumi Takabatake; Hotaka Kawai; Hitoshi Nagatsuka; Yoshihiko Furuki
Journal:  Sci Rep       Date:  2022-04-12       Impact factor: 4.379

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.