Literature DB >> 32782269

Personal identification with orthopantomography using simple convolutional neural networks: a preliminary study.

Shinpei Matsuda1, Takashi Miyamoto2, Hitoshi Yoshimura3, Tatsuhito Hasegawa2.   

Abstract

Forensic dental examination has played an important role in personal identification (PI). However, PI has essentially been based on traditional visual comparisons of ante- and postmortem dental records and radiographs, and there is no globally accepted PI method based on digital technology. Although many effective image recognition models have been developed, they have been underutilized in forensic odontology. The aim of this study was to verify the usefulness of PI with paired orthopantomographs obtained in a relatively short period using convolutional neural network (CNN) technologies. Thirty pairs of orthopantomographs obtained on different days were analyzed in terms of the accuracy of dental PI based on six well-known CNN architectures: VGG16, ResNet50, Inception-v3, InceptionResNet-v2, Xception, and MobileNet-v2. Each model was trained and tested using paired orthopantomographs, and pretraining and fine-tuning transfer learning methods were validated. Higher validation accuracy was achieved with fine-tuning than with pretraining, and each architecture showed a detection accuracy of 80.0% or more. The VGG16 model achieved the highest accuracy (100.0%) with pretraining and with fine-tuning. This study demonstrated the usefulness of CNN for PI using small numbers of orthopantomographic images, and it also showed that VGG16 was the most useful of the six tested CNN architectures.

Entities:  

Mesh:

Year:  2020        PMID: 32782269      PMCID: PMC7419525          DOI: 10.1038/s41598-020-70474-4

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Dental examination findings are unique among individuals because they consist of a combination of decayed, filled, prosthetic, and missing teeth[1]. Additionally, teeth are harder and more resistant to heat than other body tissues, including bones. Therefore, forensic dental examination has played an important role in personal identification (PI), even in large-scale accidents and disasters. Digitalization in medical and dental examinations and treatments has progressed, providing clinicians with great support in diagnoses and surgical procedures[2-4]. However, in forensic dental examination, PI has essentially been based on traditional visual comparisons of the antemortem dental records and radiographs with those obtained by postmortem dental examinations[5]. Moreover, such traditional forensic dental examinations have the disadvantage of lacking widely accepted reference points between antemortem records and postmortem examinations[1]. Finally, analyses of traditional written dental records are subjective[6]. Therefore, traditional forensic dental examinations are vulnerable to oversights and/or mistakes in PI[6]. The authors believe that one way to improve on these traditional dental examinations for dental PI that are currently used around the world would be to digitize and standardize them. Although forensic odontology with digital technologies has progressed, there is no globally accepted PI method based on digital technology due to various factors, such as special equipment and costs. With the progression of digital technologies, global digitization of forensic odontology should be promoted. In the future, such technologies will make a significant social contribution to PI in the wake of large-scale accidents and disasters. In a position statement published by the Institute of Electrical and Electronics Engineers, the following definition of artificial intelligence (AI) is suggested: AI is that activity devoted to making machines intelligent, and intelligence is that quality that enables an entity to function appropriately and with foresight in its environment[7]. The development of AI technologies has made great strides along with the growth and evolution of personal computers. At present, there is a possibility that machines will displace humans in various fields, including medicine[8]. AI technology includes deep learning, which improves a machine’s ability to represent data by deepening the layered structure of neural networks; a convolutional neural network (CNN) is a type of network for deep learning[9-15]. The authors believe that CNNs may be useful for forensic PI because they have high recognition accuracy, mainly in the field of image recognition[9]. Thus, a convergence between dental imaging examination and CNNs in the forensic field may have great potential to bring social benefits by improving preparedness for large-scale accidents and disasters due to the frequency of imaging examinations in clinical dentistry. The aim of this preliminary study was to evaluate the usefulness of PI with orthopantomography using the following six CNN architectures: VGG16, ResNet50, Inception-v3, InceptionResNet-v2, Xception, and MobileNet-v2[10-15].

Results

Participants and materials

Thirty participants (14 male and 16 female) were included in this study (Table 1). Thus, 30 pairs of orthopantomographs, for a total of 60 orthopantomographs, were analyzed in this study. The youngest participant was 15 years old, and the oldest was 86 years old. The mean ± standard deviation was 31.00 ± 15.21 years old. Twenty-nine participants underwent orthopantomography before and after tooth extractions, and one participant underwent orthopantomography twice in follow-up examinations (Table 1: No. 11). The mean ± standard deviation of the number of extracted teeth among the thirty participants was 2.03 ± 1.07, and the most common extracted tooth was the mandibular third molar. The mean and standard deviation of the number of teeth “before orthopantomography” was 29.00 ± 6.34, and that “after orthopantomography” was 26.97 ± 6.08. The mean ± standard deviation of the interval between “before orthopantomography” and “after orthopantomography” was 64.03 ± 70.29 days.
Table 1

Characteristics of participants in the study.

NoGenderAgeaTooth extraction sitesNumber of tooth extractionsNumber of teeth before extractionNumber of teeth after extractionInterval between initial and final panoramic radiography (days)
1F33381282734
2F2938, 482302849
3M2838, 482302820
4F3218, 28, 38, 4843228309
5F2518, 28, 38, 484322859
6F2938, 482302884
7F2138, 482323014
8M1838132315
9M2438, 482302835
10F212813130119
11F33Not applicable02727199
12F3338, 482292741
13M1538, 482323078
14F3628, 38, 483312849
15M2718, 482312935
16M2528, 38, 483312889
17F63481131232
18F2018, 28, 38, 4843026149
19F28381313021
20F2438, 482323021
21M8638110188
22M2938132317
23M27481302913
24M3418, 17, 27331284
25M2518, 38, 483312840
26F64381302915
27F1838, 482323030
28M4218, 28, 38, 4843127134
29M2138, 482312936
30M2038, 482272512
Average31.002.0329.0026.9764.03
Standard deviation15.211.076.346.0870.29

aTooth numbering system proposed by the Fédération dentaire internationale (FDI).

Characteristics of participants in the study. aTooth numbering system proposed by the Fédération dentaire internationale (FDI).

Training and testing a CNN model

In the training phase, the CNN used a training dataset composed of input images and their correct personal identifiers to adapt the weights (filters) of the network. After training the model, the CNN estimated the correct label from unlabeled image X′ in the testing phase. Finally, the estimated label and the correct label y were compared to calculate evaluation indexes such as accuracy, precision, and recall. In this study, the authors used accuracy as an evaluation index. The authors identified a suitable CNN architecture and training parameters for this task through comparison experiments. Each model was trained using the “before orthopantomography” dataset for 300 epochs and tested using the “after orthopantomography” dataset. In this study, the authors validated six types of CNN architecture: VGG16, ResNet50, Inception-v3, InceptionResNet-v2, Xception, and MobileNet-v2. These architectures were implemented in Keras (https://keras.io/), which is a Python deep learning library. The parameters and original papers are found in the official Keras documentation (https://keras.io/applications/#applications). The authors also validated some optimizers: SGD, RMSProp, Adadelta, Adagrad, and Adam, with a learning rate of 10−5 to 10−3. The batch size was 10. Furthermore, the authors validated two types of transfer learning methods: pretraining and fine-tuning. In pretraining, all weights in the model were trained in advance using a large-scale dataset for other tasks, and some of the weights were trained using the target dataset (e.g., the dental PI dataset in this study). The authors trained only the weights of the final layer after pretraining. In fine-tuning, all weights in the model were retrained using the target dataset after pretraining. Both transfer learning methods used the ImageNet dataset, which is a very large-scale image recognition dataset including over 14 million labeled images. In total, the authors validated 180 training patterns (6 models × 5 optimizers × 3 learning rates × 2 transfer methods).

Evaluation results

The values in the table represent the validation accuracies in the final epoch (Table 2). The authors selected the maximum accuracies from all optimizers for each method because the best optimizer and learning rate differ for each CNN architecture. In this evaluation, the VGG16 model achieved the highest accuracy, at 100.0% (Fig. 1). When drawing comparisons in transfer learning, the most accurate method to use is fine-tuning. Under fine-tuning, some CNN architectures also achieved greater than 90.0% accuracy. VGG16, despite having a relatively shallow layer structure, had the highest estimation accuracy because this dataset was very small.
Table 2

Validation accuracies for each CNN architecture and transfer learning method.

VGG16 (%)ResNet50 (%)Inception-v3 (%)Inception ResNet-v2 (%)Xception (%)MobileNet-v2 (%)
Pretraining100.03.320.026.720.016.7
Fine-tuning100.093.383.396.780.080.0
Figure 1

The loss (blue line) and accuracy (red line) in training. The x-axis measures epochs, the left y-axis measures loss, and the right y-axis measures accuracy (%). Focusing on the red line, the training and testing accuracies converge after 30 epochs. Although the training data are perfectly classified, the testing data are sometimes misclassified.

Validation accuracies for each CNN architecture and transfer learning method. The loss (blue line) and accuracy (red line) in training. The x-axis measures epochs, the left y-axis measures loss, and the right y-axis measures accuracy (%). Focusing on the red line, the training and testing accuracies converge after 30 epochs. Although the training data are perfectly classified, the testing data are sometimes misclassified.

Discussion

In this study, the authors evaluated the usefulness of the application of CNN technology in PI based on orthopantomography by validating the identification accuracy of six different CNN architectures. The results demonstrated that VGG16 was the most useful CNN architecture for PI using orthopantomography. Furthermore, VGG16 pretraining and fine-tuning using the ImageNet dataset achieved 100% identification accuracy. In cases of forensic medicine and forensic dentistry such as fire and murder victims, excluding large-scale disasters, it is important to identify individuals from a small number of people. To resolve this problem, this study indicated the possibility of accurate PI from orthopantomography image by using CNNs. According to the PubMed database, studies of CNN application have been reported since 2013 in various medical including orthopedics, oncology, ophthalmology, and neurosurgery[16-19]. In the dental field, Miki et al. first reported the application of CNNs with cone-beam computed tomography in 2017[20]. In recent years, applications of CNNs in cariology, periodontology, and endodontics have been reported, and applications in real dental clinical experience may be realized in the near future[21-23]. Tuzoff et al. reported a method of tooth detection and numbering with orthopantomography using simple CNNs, which could help save time and improve the process of filling out dental charts[24]. Kahaki et al. tried to establish an age estimation method based on global fuzzy segmentation and local feature extraction using a projection-based feature transform and a designed deep CNN model; the molars were isolated from 456 young participants' orthopantomography as an analysis target[25]. Schwendicke et al. performed a scoping review of CNNs for dental image diagnostics and concluded that CNNs may be used in diagnostic-assistance systems in the dental field[26]. At present, CNN technology is not easy for dentists to implement. In the future, generalization of such technology will further advance through the development of simple software. As a limitation of this study, the authors must note that the mean ± standard deviation of the periods between the date obtained “before orthopantomography” and the date obtained “after orthopantomography” was 64.03 ± 70.29 days; in other words, the paired orthopantomographs under comparison were obtained at relatively short intervals. This condition was due to the inclusion criteria for the study participants to rule out age-related oral changes and oral diseases such as periodontal disease and caries. Twenty-nine of the 30 participants were examined orthopantomographically before and after wisdom tooth extraction, while one participant did not undergo tooth extraction during the study. The authors considered that the reason why the VGG16 model achieved 100.0% accuracy was because the images taken as “before orthopantomography” and “after orthopantomography” were obtained under almost the same conditions and were very similar. Therefore, the results of this preliminary experiment verified the applicability of PI in an ideal environment with pairs of very similar images indicated that CNNs can classify individual orthopantomographs according to which patients they depict. In addition, because the authors were able to use only two datasets at this time, they established only a training set and a validation set. The authors have been able to perform only limited verification. Normally, it is necessary to prepare the test data. Additionally, it is necessary to train the model with a training set, perform parameter tuning with a validation set, and then verify the final estimation accuracy with a test set. This is one of the limitations of this study. In addition, the authors understand that the application of deep neural networks to small datasets may lead to the problem of overfitting. On the other hand, there are several issues with long-term changes, including periodontal bone loss, dental caries, and prosthetic rehabilitation, in studies of PI based on orthopantomography, as mentioned above. Additionally, experimental orthopantomography examination will lead to radiation exposure and ethical issues. In this study, participants who underwent orthopantomography for clinical treatment in a relatively short period were collected retrospectively and analyzed. Therefore, it was difficult to construct a large-scale image dataset. To use CNN technology for orthopantomography in preparation for large-scale accidents and disasters in the future, further validation of the number of paired orthopantomographs and the interval periods between “before orthopantomography” and “after orthopantomography” is required. Additionally, the main cause of death might be facial trauma with tooth or bone fractures. Therefore, various clinical situations such as changes in maxillofacial and dental status should be considered in PI, and PI should be performed with high accuracy. Based on this preliminary study, it is important to proceed with PI research using CNN technology but also considering the cause of death in order to apply the results of this study to actual forensic cases in the future. Regarding the loss function, the best-known loss function for classification tasks is cross-entropy loss, which is defined by the equation in this study. Mean squared error and mean absolute error can also be used as loss functions in classification, and there are other loss functions as well, such as focal loss, class-balanced loss, and triplet loss. Because this experiment was a preliminary study, the authors adopted the best-known loss function: cross-entropy loss. Generally, orthopantomography requires standing and/or proper posture. Therefore, obtaining postmortem orthopantomographs may be difficult. Franco et al. suggested that postmortem full-body CT images could become a valuable tool in individual dental identification procedures[27]. Ohtani et al. reported radiographic imaging examination using portable devices acceptable for prone position[28]. In the future, the authors suggest that dentists and engineers should consider methods of constructing an orthopantomography-like image that is acceptable for postmortem examination by using data from another modality, such as CT images; CNN technology could then be applied for PI with those images. In conclusion, this study demonstrated the usefulness of CNN for PI using a small number of orthopantomographic images and that VGG16 was more useful than five other CNN architectures. Further validation in the number of paired images and in the interval between “before orthopantomography” and “after orthopantomography” is required to use this technology for large-scale accidents and disasters in the future.

Materials and methods

The authors designed a retrospective study. The participants were patients who received orthopantomography for clinical treatment at the University of Fukui Hospital twice from January 2018 to September 2019 and at relatively short intervals to rule out age-related oral changes and oral diseases such as periodontal disease and caries. Paired orthopantomographs of those participants were analyzed for the development of the PI method. In this study, based on the examination date, paired orthopantomographs were defined as “before orthopantomography” and “after orthopantomography” (Fig. 2). The orthopantomography data used in this study are stored in the Image Management System of the University of Fukui Hospital, and the details of the participants are shown in Table 1.
Figure 2

The design of this study was based on paired orthopantomographs analyzed with six convolutional neural network architectures. The paired orthopantomographs of participant No. 1 are shown as (A) “before orthopantomography” and (B) “after orthopantomography”, and tooth extraction sites are indicated by white arrows in this figure.

The design of this study was based on paired orthopantomographs analyzed with six convolutional neural network architectures. The paired orthopantomographs of participant No. 1 are shown as (A) “before orthopantomography” and (B) “after orthopantomography”, and tooth extraction sites are indicated by white arrows in this figure.

Simple convolutional neural networks

Machine learning is a method to explore a function f(⋅) that can convert x to y from a dataset consisting of input data x with the correct label y. The goal is to optimize the formula y = f(x) from a given dataset. A CNN is a neural network model composed of many layers, including a convolution layer and a pooling layer. CNNs are aggressively applied in the image recognition field—for example, cancer detection from X-ray images. The basic principle of the convolution process is to calculate the convolution from a raw image with filters. Filters, which are similar to weights in a neural network, are adoptively adjusted in the training process. By this process, a computer aims to acquire the effective filters to detect the target from the dataset. The basic principle of the pooling process is to calculate the representative value from fixed-size windows—for example, a 2 × 2 max pooling layer calculates the max values from each 2 × 2 array of pixels. Through this process, it is said that a computer acquires a robust feature representation of the target position.

Formulation of this problem

The authors formulated a dental PI task to complete via a CNN model. In this task, a dental image is given, and the personal identifier is required as an output. The input for machine learning is a dental image X, and the output is a personal identifier y. Therefore, CNN model f(⋅) takes the role of a converter that converts input image X into personal identifier y. This is a simple classification task represented by the formula y = f(X). In order to accurately estimate the personal identifier y, CNN filters are adapted based on the loss function. There are many CNN architectures and loss functions in the facial recognition field; however, as a preliminary experiment, this study adopted a simple CNN classifier with a common loss function called categorical cross-entropy loss (L). L is represented by the following formula:where N is the number of target people, t is a correct label (one-hot vector), and f(X) is the i-th CNN’s output. CNN adopts its filters to minimize L in the training phase.

Ethics approval and informed consent

This study was approved by the Institutional Research Board (Ethical Committee of the University of Fukui, Faculty of Medical Sciences, No. 20190043). There were no ethical issues in conducting this study because it was a retrospective study targeting examination images obtained for clinical treatment, and the authors declared that all methods were performed in accordance with the relevant guidelines and regulations (Declaration of Helsinki). Additionally, informed consent was obtained from all subjects or, if subjects were under 18, from a parent and/or legal guardian.
  18 in total

1.  Deep Learning for the Radiographic Detection of Apical Lesions.

Authors:  Thomas Ekert; Joachim Krois; Leonie Meinhold; Karim Elhennawy; Ramy Emara; Tatiana Golla; Falk Schwendicke
Journal:  J Endod       Date:  2019-06-01       Impact factor: 4.171

2.  Classification of teeth in cone-beam CT using deep convolutional neural network.

Authors:  Yuma Miki; Chisako Muramatsu; Tatsuro Hayashi; Xiangrong Zhou; Takeshi Hara; Akitoshi Katsumata; Hiroshi Fujita
Journal:  Comput Biol Med       Date:  2016-11-12       Impact factor: 4.589

3.  Feasibility and validation of virtual autopsy for dental identification using the Interpol dental codes.

Authors:  Ademir Franco; Patrick Thevissen; Walter Coudyzer; Wim Develter; Wim Van de Voorde; Raymond Oyen; Dirk Vandermeulen; Reinhilde Jacobs; Guy Willems
Journal:  J Forensic Leg Med       Date:  2012-10-10       Impact factor: 1.614

4.  Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network.

Authors:  Adhish Prasoon; Kersten Petersen; Christian Igel; François Lauze; Erik Dam; Mads Nielsen
Journal:  Med Image Comput Comput Assist Interv       Date:  2013

Review 5.  Convolutional neural networks for dental image diagnostics: A scoping review.

Authors:  Falk Schwendicke; Tatiana Golla; Martin Dreher; Joachim Krois
Journal:  J Dent       Date:  2019-11-05       Impact factor: 4.379

6.  How Artificial Intelligence Will Change Medicine.

Authors:  Claudia Wallis
Journal:  Nature       Date:  2019-12       Impact factor: 49.962

Review 7.  Why a dentist for identification?

Authors:  D Sweet
Journal:  Dent Clin North Am       Date:  2001-04

8.  Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm.

Authors:  Jae-Hong Lee; Do-Hyung Kim; Seong-Nyum Jeong; Seong-Ho Choi
Journal:  J Dent       Date:  2018-07-26       Impact factor: 4.379

9.  Tooth detection and numbering in panoramic radiographs using convolutional neural networks.

Authors:  Dmitry V Tuzoff; Lyudmila N Tuzova; Michael M Bornstein; Alexey S Krasnov; Max A Kharchenko; Sergey I Nikolenko; Mikhail M Sveshnikov; Georgiy B Bednenko
Journal:  Dentomaxillofac Radiol       Date:  2019-03-05       Impact factor: 2.419

10.  Application of a real-time 3-dimensional navigation system for treatment of synovial chondromatosis of the temporomandibular joint: A case report.

Authors:  Shinpei Matsuda; Hitoshi Yoshimura; Kazuo Sano
Journal:  Medicine (Baltimore)       Date:  2019-05       Impact factor: 1.817

View more
  6 in total

1.  A fully automated method of human identification based on dental panoramic radiographs using a convolutional neural network.

Authors:  Young Hyun Kim; Eun-Gyu Ha; Kug Jin Jeon; Chena Lee; Sang-Sun Han
Journal:  Dentomaxillofac Radiol       Date:  2021-12-02       Impact factor: 3.525

2.  Current applications and development of artificial intelligence for digital dental radiography.

Authors:  Ramadhan Hardani Putra; Chiaki Doi; Nobuhiro Yoda; Eha Renwi Astuti; Keiichi Sasaki
Journal:  Dentomaxillofac Radiol       Date:  2021-07-08       Impact factor: 2.419

Review 3.  Personal identification with artificial intelligence under COVID-19 crisis: a scoping review.

Authors:  Shinpei Matsuda; Hitoshi Yoshimura
Journal:  Syst Rev       Date:  2022-01-06

4.  Lingual bone thickness in the apical region of the horizontal mandibular third molar: A cross-sectional study in young Japanese.

Authors:  Shinpei Matsuda; Hitoshi Yoshimura
Journal:  PLoS One       Date:  2022-01-25       Impact factor: 3.240

Review 5.  Possibilities and challenges in digital personal identification using teledentistry based on integration of telecommunication and dental information: a narrative review.

Authors:  Shinpei Matsuda; Hitoshi Yoshimura
Journal:  J Int Med Res       Date:  2022-04       Impact factor: 1.573

6.  A pilot study of an automated personal identification process: Applying machine learning to panoramic radiographs.

Authors:  Adrielly Garcia Ortiz; Gustavo Hermes Soares; Gabriela Cauduro da Rosa; Maria Gabriela Haye Biazevic; Edgard Michel-Crosato
Journal:  Imaging Sci Dent       Date:  2021-05-06
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.