| Literature DB >> 35607622 |
Hu Chen1, Che Sun1, Peixi Liao2, Yancun Lai1, Fei Fan3, Yi Lin1, Zhenhua Deng3, Yi Zhang1.
Abstract
When accidents occur, panoramic dental images play a significant role in identifying unknown bodies. In recent years, deep neural networks have been applied to address this task. However, while tooth contours are significant in classical methods, few studies using deep learning methods devise an architecture specifically to introduce tooth contours into their models. Since fine-grained image identification aims to distinguish subordinate categories by specific parts, we devise a fine-grained human identification model that leverages the distribution of tooth masks to distinguish different individuals with local and subtle differences in their teeth. First, a bilateral branched architecture is designed, of which one branch was designed as the image feature extractor, while the other was the mask feature extractor. In this step, the mask feature interacts with the extracted image feature to perform elementwise reweighting. Additionally, an improved attention mechanism was used to make our model concentrate more on informative positions. Furthermore, we improved the ArcFace loss by adding a learnable parameter to increase the loss of those hard samples, thereby exploiting the potential of our loss function. Our model was tested on a large dataset consisting of 23,715 panoramic X-ray dental images with tooth masks from 10,113 patients, achieving an average rank-1 accuracy of 88.62% and rank-10 accuracy of 96.16%.Entities:
Keywords: human identification; neural network; panoramic dental images; tooth contour
Year: 2022 PMID: 35607622 PMCID: PMC9122963 DOI: 10.1016/j.patter.2022.100485
Source DB: PubMed Journal: Patterns (N Y) ISSN: 2666-3899
Figure 1An overview of our network
Figure 2Flowcharts
(A) Model architecture in basic branch. (B) Model architecture in mask branch.
Figure 3Similar radiographs of two different individuals, who have lost the same teeth at the same place and had new ones implanted
Figure 4Architecture of channel attention blocks
(A) SE block. (B) BPCD block. (C) GC block. (D) Proposed GC-BPCD block.
Figure 5Heatmaps
(A) is the original image, and (B) and (C) are heatmaps generated by different models using (A). (B) is generated by GCNet, (C) is generated by our poposed model.
SD values
| Module | GC block | Proposed model |
|---|---|---|
| Images | 7.12 | 11.18 |
Figure 7Loss curves of three different loss functions
Best viewed in color.
With or without masks
| Input | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| Images | 34.47 | 57.63 | 26.39 |
| Images + masks | 38.94 | 63.52 | 31.22 |
Figure 6Softmax loss curve
Best viewed in color.
Different loss functions
| Loss function | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| L-softmax loss | 44.94 | 67.52 | 38.22 |
| A-softmax loss | 38.57 | 63.25 | 27.11 |
| Cosine loss | 47.46 | 71.93 | 38.77 |
| ArcFace loss | 56.58 | 79.47 | 47.91 |
| Curricular loss | 45.79 | 68.60 | 38.19 |
| Dynamic ArcFace loss | 65.96 | 85.26 | 57.10 |
Attention modules
| Attention module | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| SENet | 51.14 | 75.53 | 41.15 |
| CBAM | 50.53 | 75.96 | 40.93 |
| LCANet | 68.60 | 89.04 | 60.14 |
| ECA-Net | 52.11 | 75.97 | 44.04 |
| GCNet | 64.74 | 85.96 | 57.63 |
| Proposed module | 75.53 | 90.87 | 67.82 |
Figure 8Grad-CAM visualizations
The first row contains three different images, and in the following rows, each row contains three heatmaps correspond to different images.
Backbones
| Backbone module | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| ResNet | 56.92 | 71.83 | 48.52 |
| U-Net | 61.30 | 74.11 | 53.38 |
| Proposed | 67.62 | 77.53 | 62.34 |
| Proposed + GN | 69.82 | 87.01 | 61.50 |
Feature fusion
| Feature fusion | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| No feature fusion | 69.82 | 87.01 | 61.50 |
| Elementwise addition | 81.14 | 93.51 | 73.95 |
| Elementwise multiplication | 87.81 | 96.67 | 82.92 |
Loss functions for mask loss
| Loss function | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| Softmax | 83.16 | 94.65 | 78.06 |
| Dynamic ArcFace loss | 86.05 | 96.23 | 81.32 |
| L1 | 87.81 | 96.67 | 82.92 |
Figure 9Loss curves of three different loss functions
Best viewed in color.
Other models
| Model | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| ResNet | 54.21 | 77.72 | 42.97 |
| ResNeXt | 58.68 | 81.14 | 50.27 |
| InceptionNet | 48.77 | 75 | 39.03 |
| DenseNet | 44.56 | 71.05 | 34.32 |
| EfficientNet | 69.21 | 87.20 | 61.35 |
| Ajaz and Kathirvelu | 17.51 | 24.39 | 8.73 |
| Oktay | 26.75 | 47.37 | 18.00 |
| DentNet | 39.47 | 65.09 | 28.17 |
| LCANet | 78.86 | 92.81 | 72.67 |
| Proposed | 87.81 | 96.67 | 82.92 |
Five-fold cross-validation
| Number | Rank-1 accuracy | Rank-10 accuracy | TAR(@FAR = 10-4) |
|---|---|---|---|
| 1 | 89.91 | 96.57 | 83.90 |
| 2 | 88.68 | 96.32 | 83.88 |
| 3 | 87.74 | 95.67 | 83.43 |
| 4 | 88.43 | 96.08 | 86.27 |
| 5 | 88.34 | 96.17 | 83.23 |
| Average | 88.62 | 96.16 | 84.14 |