Literature DB >> 35607622

A fine-grained network for human identification using panoramic dental images.

Hu Chen1, Che Sun1, Peixi Liao2, Yancun Lai1, Fei Fan3, Yi Lin1, Zhenhua Deng3, Yi Zhang1.   

Abstract

When accidents occur, panoramic dental images play a significant role in identifying unknown bodies. In recent years, deep neural networks have been applied to address this task. However, while tooth contours are significant in classical methods, few studies using deep learning methods devise an architecture specifically to introduce tooth contours into their models. Since fine-grained image identification aims to distinguish subordinate categories by specific parts, we devise a fine-grained human identification model that leverages the distribution of tooth masks to distinguish different individuals with local and subtle differences in their teeth. First, a bilateral branched architecture is designed, of which one branch was designed as the image feature extractor, while the other was the mask feature extractor. In this step, the mask feature interacts with the extracted image feature to perform elementwise reweighting. Additionally, an improved attention mechanism was used to make our model concentrate more on informative positions. Furthermore, we improved the ArcFace loss by adding a learnable parameter to increase the loss of those hard samples, thereby exploiting the potential of our loss function. Our model was tested on a large dataset consisting of 23,715 panoramic X-ray dental images with tooth masks from 10,113 patients, achieving an average rank-1 accuracy of 88.62% and rank-10 accuracy of 96.16%.
© 2022 The Author(s).

Entities:  

Keywords:  human identification; neural network; panoramic dental images; tooth contour

Year:  2022        PMID: 35607622      PMCID: PMC9122963          DOI: 10.1016/j.patter.2022.100485

Source DB:  PubMed          Journal:  Patterns (N Y)        ISSN: 2666-3899


Introduction

The identification of individuals plays an important role in forensics. Several biometric features, including DNA, fingerprints, faces, and voices, have been used in human identification. However, these features are susceptible to decay in the natural environment. In contrast, teeth are covered by enamel, which is the hardest tissue in the human body. Teeth do not undergo major morphological changes after death, which makes teeth effective features in matching antemortem and postmortem records. Digital images are commonly used in human identification due to their acceptable costs and accuracy. Early on, antemortem and postmortem records were manually compared by experts using dental restoration and tooth morphology,, which incurred significant time costs with subjective errors. Then, automated comparison systems were developed, and classical methods were adopted. Jain and Chen used a semiautomatic contour extraction method to address the problem of fuzzy tooth contours and then retrieved the best match to a given postmortem radiograph from a database of 100 images. Nomir and Abdel-Mottaleb used integral projection to separate teeth, and the signatures of vectors of tooth contours were compared. In recent years, deep neural networks have been applied in this area and have made substantial progress. Fan et al. built a human identification system with a convolutional neural network using panoramic dental radiographs. Lai et al. presented a network that incorporated a channel attention module to selectively emphasize channel information. Sathya and Neelaveni used transfer learning and classified teeth into four classes, thereby increasing the performance of the model. Wu et al. proposed a multisupervision network with an attention-based mechanism to obtain an attention mask for human identification. These previous studies substantially improved the accuracy and reliability of human identification using dental images. However, these deep-learning-based works used entire images as input and processed the textural feature together with the contour feature without distinction. Tooth contours have not been treated in much detail but have been proved to be meaningful features in previous methods.9, 10, 11 Fine-grained image classification refers to problems of distinguishing subordinate categories among basic categories. In this paper, we treat the human identification problem as a fine-grained classification problem since the targets are the same. Moreover, fine-grained classification requires the model to focus on areas with different details but that are similar overall; in our model, we adopted several methods to distinguish fine-grained categories. First, our model is composed of a basic branch and a mask branch to pay strong attention at the part level. The basic branch extracts the contextual deep visual features from the whole image, while the mask branch processes the masks (i.e., segmented tooth contours) to obtain the mask features. The mask features cooperate with the contextual features to generate the weighted features. The mask features can be seen as weights of the basic features, similar to convolutional filters in dynamic filter networks. They enhance the correlation between bilateral networks, guide the model to pay strong attention to the fine-grained tooth contours, and adjust the training procedure automatically. Second, as a supplement to the strong attention mentioned above, to learn discriminative parts inside and outside of masks automatically, an improved attention mechanism is contained in the basic branch. The mechanism can also ease the constraint of the local receptive field brought by the convolution operation. Furthermore, to distinguish similar individuals (hard samples), inspired by LCANet, we improved the ArcFace loss and propose an improved loss function (dynamic ArcFace loss) as the loss function of the basic features and weight features. The main contributions of our work are summarized as follows: (1) we propose a bilateral network model based on masks and feature fusion to solve the individual identification problem using dental panoramic X-ray images. The network architecture of this model is uncomplicated, and a large number of experiments have proved that it achieves better results than previous methods. (2) An attention module is devised that is simple but effective and can be easily used in other existing models without modifying their backbones. (3) We present an improved loss function (the dynamic ArcFace loss) that adjusts the loss in the training process using learnable parameters and generates more discriminative features. Our conclusions are confirmed by further experiments. On the test set, which is composed of 1,000 individuals, our model achieves a rank-1 accuracy of 87.81% and a rank-10 accuracy of 96.67%, and the accuracy is 82.92% when the false acceptance rate (FAR) is one in 10,000. The average rank-1 and rank-10 accuracy rates of the 5-fold cross-validation experiment are 88.62% and 96.16%, respectively.

Results and discussion

Related work

Fine-grained models

Fine-grained classification models are designed to distinguish between multiple subcategories, among which the differences can be subtle. Part-based R-CNNs extract features from whole-object and part detectors and handle classification issues using methods borrowed from object detection. Wei et al. proposed a mask-CNN model for aggregating contextual and part attributes based on part-annotated fine-grained images. Yang et al. proposed a self-supervision mechanism consisting of a navigator network, a teacher network, and a scrutinizer network to localize informative regions automatically while optimizing these networks in a pipeline. Our bilateral network infuses information derived from the tooth contour mask into the main contextual feature and trains branches together.

Attention modules

By using attention modules, networks are able to determine long-range dependencies and examine more meaningful areas. The global attention mechanism proposed in the nonlocal network contains a self-attention module that models multilevel global dependency relationships. GCNet soon followed. GCNet combines the channel attention mechanism (SE block) in SENet and the global attention mechanism in the nonlocal network to propose a global attention mechanism global context (GC) block that derives contextual information to achieve global dependency modeling. Built on the SE block, the channel attention module in LCANet introduces a structure composed of BN-PreLU-Conv1∗1-Dropout (LCANet module), which keeps the channel interaction in a proper range and promotes convergence. In our model, an LCANet model is used in the lateral branch of the GC block, resulting in an enhanced learning capability.

Cosine loss function

Solving the human identification problem on a large dataset deeply relies on devising a proper loss function. Early biometric vision tasks used the softmax loss as the loss function of the model. Subsequently, improved loss functions based on the softmax loss function to better distinguish features have been proposed. Among these improvements, Wen et al. proposed the center loss to learn the intermediate features of each individual to reduce the intraclass gap, and Kemelmacher-Shlizerman et al. proposed the L-softmax loss by adding angular constraints to the features to improve the discernibility of features. Subsequently, Liu et al. improved the accuracy of the L-softmax loss on the open set recognition task by normalizing the weight parameters and proposed the A-softmax loss. Wang et al. proposed a cosine loss to improve the softmax loss. The cosine loss adjusted the features and weights through L2 normalization and added angular constraints to features. Then, the ArcFace loss proposed by Deng et al. improved the cosine loss and changed the angular constraint of the feature from the cosine value to the angular value comparison so as to reduce the intraclass gap and expand the interclass gap. Our improved loss function adjusts the weights automatically to better learn the distribution of hard samples.

Methods

Bilateral-branch network

In this part, we illustrate a bilateral-branch network model, which pays strong attention to tooth contours. The basic branch extracts the basic features, and the mask branch extracts mask features to enhance the identification from a more fine-grained perspective. The framework is shown in Figure 1.
Figure 1

An overview of our network

An overview of our network

Basic branch

To excavate the features concerning tooth morphology from the mask, we combine the original tooth image and its corresponding tooth segmentation map (i.e., mask) using the channel axis and use the combination as the input of the basic network. Then, the model extracts the contextual features from the original image and local tooth contour features from the mask. Note that unless otherwise specified, the convolutional layer in the basic branch is composed of a normal convolutional layer, batch normalization (BN), and an activation function (parametric rectified linear unit [PreLU]). After being sent to the network, the input first goes through a convolutional layer with a 3 × 3 convolutional kernel and 64 channels and then is downsampled by maximum pooling. Next, it goes through four convolutional modules (I, II, III, IV). Each module is composed of BN and two convolutional blocks (each convolutional block is composed of two convolution layers with a 3 × 3 kernel size), and the numbers of channels of the blocks are 64, 128, 256, and 512 in sequence. In addition, after the convolution modules (II, III, IV), a convolutional layer with a step size of 2 and a 1 × 1 convolutional kernel is added to complete the downsampling operation (only a normal convolutional layer and BN are used in the downsampling convolutional layer). Due to the local receptive field of the convolutional neural network, the deeper the network is, the more features fade. To alleviate the harm caused by the local receptive field, two methods were adopted to enhance feature propagation. First, each convolutional block in the above four convolution modules adopts a head-to-tail feature fusion strategy to enhance feature propagation and reuse parameters. In addition, at the end of each convolutional block, an improved attention mechanism is included. It models the global contextual dependencies in the feature map better and strengthens the ability to obtain key information in the feature map. Subsequently, a structure composed of BN and dropout is added at the end of all convolutional layers to adjust the distribution of features. Finally, a fully connected layer is used to extract the 1,024-dimensional deep visual features of teeth. The flowchart is shown in Figure 2A.
Figure 2

Flowcharts

(A) Model architecture in basic branch. (B) Model architecture in mask branch.

Flowcharts (A) Model architecture in basic branch. (B) Model architecture in mask branch.

Mask branch

Due to the importance of tooth contours in classical human identification methods,, we devise a specific mask branch to make full use of the morphological characteristics of teeth. Given a tooth segmentation image (i.e., mask) determined by convolutional neural network (CNN) corresponding to a panoramic dental image, there are only two different pixel values in the segmentation image: a value of one for the pixels in the target area of the tooth, and a value of zero for the background area. That binary distribution strengthens the mask branch such that it can capture the slight feature distribution changes in different masks accurately. After passing to the mask branch, the mask image goes through four convolutional blocks, and each convolutional block is composed of two convolutional layers with a 3 × 3 convolutional kernel. Each convolutional layer is composed of a normal convolutional layer, a group normalization (GN) layer and an activation function (PReLU) structure. The numbers of channels of the four convolutional blocks are 64, 128, 256, and 512, respectively. A 1 × 1 convolutional kernel with a step size of 2 is set after each convolution block to downsample the feature map. Finally, two fully connected layers are used: one obtains the 1,024-dimensional mask features and interacts with the basic features, and the other predicts the class (mask logit) and calculates the mask loss directly. Our branch network replaces traditional BN with GN and utilizes the trait that GN adjusts the data distribution within the channel group to improve the characterization capabilities of mask features. The flowchart is shown in Figure 2B.

Feature fusion

In dental panoramic X-ray images, the background area is much larger than the target area; additionally, due to different formal clinical treatments, many individuals have similar tooth characteristics. As shown in Figure 3, different individuals have lost the same teeth and have had new teeth implanted, which easily leads to a higher similarity of the characteristics excavated by the model. In the last subsection, we devised a mask branch that extracts the mask features to better collect the morphological information of teeth. Until now, our branch networks have been two separate units. To strengthen the connection between the two branches, motivated by Kumar et al., we adopt an elementwise feature fusion strategy. This strategy performs elementwise multiplication on the features extracted by the basic branch and mask branch. Furthermore, the human identification problem can be treated as a fine-grained classification problem with small samples. This feature fusion strategy can help the network discover more discriminative visual features from tooth images in each subclass (one person) and overcomes the existing challenging small-sample problem. The practical significance of each part of the model will be proved by ablation studies in the experimental section.
Figure 3

Similar radiographs of two different individuals, who have lost the same teeth at the same place and had new ones implanted

Similar radiographs of two different individuals, who have lost the same teeth at the same place and had new ones implanted

GC-BN-PReLu-Conv1x1-Dropout attention module

The attention mechanism has become a feasible approach to fine-grained image recognition. Since directly repeating convolution layers would lead to ineffective modeling of the long-range dependency, GCNet proposes the GC block (as shown in Figure 4C) to model the global context. Specifically, the GC block utilizes a context modeling module (as shown in the red box of Figure 4C) which contains a Conv 1∗1 and a matrix multiplication to form a global context feature. However, the GC block projects channel features into a one-dimensional space, which may omit information in other dimensions28, 29, 30 and cause weak global interactions among the global context features in the human identification task with dental images. As shown in Figure 5B, we visualized the attention map (heatmap) of the GC block for different positions. It was observed that the responses in the surrounding background and the target tooth region were analogous and salient, indicating that global context features are indiscriminative for different positions. To statistically verify this observation, we analyzed the SD value of attention weights in different positions in Table 1) and determined that the value is small, indicating that attention weights learned by the GC block are similar for different positions, which further verifies that the GC block is inefficient in modeling the long-range dependencies.
Figure 4

Architecture of channel attention blocks

(A) SE block. (B) BPCD block. (C) GC block. (D) Proposed GC-BPCD block.

Figure 5

Heatmaps

(A) is the original image, and (B) and (C) are heatmaps generated by different models using (A). (B) is generated by GCNet, (C) is generated by our poposed model.

Table 1

SD values

ModuleGC blockProposed model
Images7.1211.18
Architecture of channel attention blocks (A) SE block. (B) BPCD block. (C) GC block. (D) Proposed GC-BPCD block. Heatmaps (A) is the original image, and (B) and (C) are heatmaps generated by different models using (A). (B) is generated by GCNet, (C) is generated by our poposed model. SD values Inspired by LCANet, in which the BN-PReLu-Conv1x1-Dropout (BPCD) block can capture interactions, we attempt to develop an efficient attention instantiation by developing the BN-PReLu-Conv1x1-Dropout module at the context modeling module to aggregate the global context features and reach a new instantiation of the general framework, which is referred to as the GC-BPCD block. Specifically, the BPCD module can be regarded as vector concatenation-based attention. Compared with the original context modeling module (as shown in Figure 4C), the BPCD module can strengthen the globe interactions and feature aggregation by repeatedly refining the global context features, producing more effective long-range dependencies. Specifically, the coefficient of the negative part is not constant, which can improve the model fitting with minimal overfitting risk. Our module is formulated aswhere denotes For Equation 1, is the input feature, where is the number of positions in the feature map. represents the parameters in Conv 1x1 before softmax; denotes LCA module(Equation 2); and denotes softmax. corresponds to the context modeling part in Figure 4D, and the remaining denotes the transform part in Figure 4D, where and are parameters of two Conv 1×1 s. denotes the residual part. For Equation 2, equation denotes the BPCD part in Figure 4B, and W is the parameter of Conv 1x1. To prove the effect of our attention module, we choose three images randomly and then calculate the SD of the channel attention values in the GC-BPCD module and GC block separately. The results of the proposed model are shown in Figure 5C and Table 1. The responses between the surrounding background and the target tooth region are different and the SD value is higher, which indicates that the proposed GC-BPCD block strengthens the modeling of the long-range dependencies and enables better recognition performance. More experiments are discussed later in the experiment section.

Dynamic ArcFace loss

In the last two subsections, part-level strong attention and soft attention have been established to focus on discriminative parts, and then to promote fine-grained feature learning. In this subsection, an improved loss is devised to improve the discernibility of features at the object level. For a better comprehension of our dynamic ArcFace loss, in this section, we first review the design of the ArcFace loss and former loss measures.

Revisiting ArcFace loss and the former

The softmax loss, known as the most classical loss function in biometric recognition, can be expressed aswhere is the feature of the ith sample of the th category and has a dimension of d, denotes the weight parameter, denotes the jth column of W, is the bias, N is the batch size, and n is number of categories. Despite its simplicity, the softmax loss fails to help the model extract unique features. The following cosine loss performs the corresponding improvement, and it is expressed aswhere is the angle between and . The cosine loss adjusts feature and weight using normalization and adds a restriction m on the angle. Subsequently, the ArcFace loss further improves the cosine loss, and the restriction on the cosine is directly changed to restriction on the angle asThis increases the interclass distance and decreases the intraclass distance. Taking a binary classification problem as an example, represents the angle between weight and feature in category . Assuming that a certain feature belongs to category 0, then the ultimate goal of the cosine loss is to make . The cosine function is monotonically decreasing in the range of , so increasing the angular constraint m forces the model to learn from the features that have larger interclass distances and smaller intraclass distances. Similarly, the ArcFace loss applies constraints directly to the angular value in order to make , which further increases the interclass distance and decreases the intraclass distance. The ArcFace loss experiment proves that the ArcFace loss improves the performance and is better than the cosine loss. However, as shown in Figure 7, our experiment reveals that there is a definite increase in the ArcFace loss that makes the training procedure more difficult, and the loss remains basically unchanged at approximately 0.1 in the late stage of training. Huang et al. show that the ArcFace loss does not pay enough attention to difficult samples, which relatively increases the loss from simple samples, thus stabilizing the loss value since the model has fully learned the distribution of simple samples.
Figure 7

Loss curves of three different loss functions

Best viewed in color.

Softmax loss curve Best viewed in color. Loss curves of three different loss functions Best viewed in color.

Dynamic ArcFace loss (improved loss)

To solve this problem, inspired by focal loss, we seek to reduce the loss produced by those easy samples. The formula of our loss function is expressed aswhere is the ArcFace loss (defined in Equation 5). In the training procedure, our dynamic ArcFace loss suppresses easy samples with large losses while promoting hard samples with small losses by increasing their loss values, thus encouraging the model to learn hard samples automatically. Different from curricular loss, the proposed dynamic ArcFace loss adjusts the weights of samples outside of the log function and is iteration agnostic. Different from focal loss, we introduce cosine similarity for human identification tasks and set γ to 2 under the objective of robustness. Different from ArcFace loss, dynamic ArcFace loss is equipped with a flexible mechanism to address the difficulty of samples. All parts of the proposed loss function are devised to repair the former drawbacks. The performance of dynamic ArcFace loss also outperforms ArcFace loss and curricular loss, as Figure 7 shows. The final loss is an aggregation of the dynamic ArcFace loss and loss. We use the dynamic ArcFace loss in the optimization of basic features and weight features . For mask loss, since the mask logit in Figure 1 is a direct prediction of individuals, we choose the loss. The final loss can be expressed aswhere λ is the hyperparameter and is set to 0.1. This is done because, at the beginning of the training procedure, fluctuates extremely and, if it is set to 1, the model will not converge. More improvements will be given in the section “experimental procedures.”

Experimental procedures

Resource availability

Lead contact

Further information and requests for resources and reagents should be directed to and will be fulfilled by the lead contact, Yi Zhang (yzhang@scu.edu.cn).

Materials availability

This study did not generate new unique reagents.

Data and code availability

We have produced and used human data (dental images) in this study, so, due to ethical consideration, we are not able to release the dataset publicly. However, interested researchers can ask the lead contact for access to the data. Codes for panoramic dental images are available at https://github.com/BreezeHavana/FGHINet. All original code has been deposited at Zenodo under the DOI 10.5281/zenodo.6223257 and is publicly available as of the date of publication. Any additional information required to reanalyze the data reported in this paper is available from the lead contact upon request.

Ethics statement

All study participants provided informed consent, and the study design was approved by an ethics review board.

Data acquisition and hyperparameters

All the panoramic X-ray dental images used in our experiments, which were taken before and after oral clinical treatment, were collected from the West China Hospital of Stomatology and the Peking University Hospital of Stomatology. The dataset contains tooth filling, root canal, and restorations. Images derived from different devices do not affect the results. Our dataset consists of 23,715 images from 10,113 patients. The dataset is randomly split into a training set that consists of 21,575 images from 9,113 patients and a test dataset composed of 2,140 images from 1,000 patients. The training set and test set are completely separated from the individual axis; that is, if an image belonging to an individual is chosen by the training set, all the images of the individual should be collected in the training set, which is a concept borrowed from human identification. For better generalization performance, we further separate the test set into a register set and a validation set. If there are M total dental images that belong to an individual, we randomly pick one for the register set, and the other images are the validation set. Our device is equipped with Nvidia Geforce 2080Ti; the cuda version is 10.2. A total of 9 s is needed to train 100 iterations, and the training time is approximately 2 h. The inference time is 0.01 s per image. For preprocessing, We first crop the image size to (2,100, 850) to exclude irrelevant areas and resize the scale from (2,100, 850) to (128, 128) as the size of input image, then we reduce the influence of artifacts in the original image using contrast enhancement, and finally we set rectangles with a size of (15, 30) in the target area randomly to perform random occlusion. This occlusion is performed to improve generalization and stability of the model. The masks in our dataset are determined by CNN. We choose stochastic gradient descent (SGD) with a momentum of 0.9 and a decay of 0.0005 as our optimizer. The batch size is 16, the dropout is 0.6, and the initial learning rate is 0.01. The learning rate decreased by half every 30,000 steps, and all training steps are completed after 80,000 steps. In the testing procedure, we only use the basic branch to extract 1,024-dimensional features, use the cosine distance as the similarity, and then rank them. More specifically, we separately extract the features of images from the register set and validation set by a trained model, choose an image from the validation set randomly, match its feature with the feature of every image in the register set, and calculate their similarities. Then, we rank all the similarities from high to low. Finally, the identity of the image is determined by the image in the register set with the highest similarity.

Ablation studies in basic branch

In this section, we perform several ablation studies to prove the importance of each part of the network. All the experiments are implemented without a mask branch or feature fusion in this part.

Masks

Concentrating on the role masks play in our model, we abandoned the other modified units mentioned above with only the basic backbone and softmax loss left and used few images and a combination of images and masks as input, respectively. As shown in Table 2, after we added masks as supplemental information, the rank-1 accuracy increases by 4.47% and the true acceptance rate (TAR) increases by 4.83% when FAR is . The results prove that adding masks improves the performance, and the accuracy is still low. Moreover, we visualize the training loss in Figure 6, which indicates that our model converges fast regardless of whether masks are added, and the main reason is that the softmax loss is easy to optimize and converges fast.
Table 2

With or without masks

InputRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
Images34.4757.6326.39
Images + masks38.9463.5231.22
Figure 6

Softmax loss curve

Best viewed in color.

With or without masks

Dynamic ArcFace loss (improved loss)

In this section, we continue to conduct experiments on the impact of our improved loss, and the combination of images and masks is taken as the input of the basic branch. Five other loss functions (L-softmax loss, A-softmax loss, cosine loss, ArcFace loss, and curricular loss) are adopted as comparisons to verify the improved loss function in this chapter. As shown in Table 3, among all loss functions, the cosine loss, ArcFace loss, and dynamic ArcFace loss have better results than other loss functions because these three losses add angular constraints to features, and it is beneficial for the model to learn features with larger interclass distances and smaller intraclass distances. Furthermore, the ArcFace loss and dynamic ArcFace loss directly affect the angle, thus achieving better results than the cosine loss. In addition, both the curricular loss and the dynamic ArcFace loss improve the ArcFace loss. The difference is that, in different training procedures, the angular constraint weights are determined by different numbers of iterations, and the weight increases as the iteration number increases, which leads to a fast decrease in the loss. However, our dynamic ArcFace loss adjusts the weight by the similarity score of the sample to reduce the loss of easily distinguishable samples, which is why the dynamic ArcFace loss performs better than the curricular loss. We also visualize the training losses of different loss functions, as shown in Figure 7. Figure 7 shows that, when the ArcFace loss is used in training, the loss curve shows a significant decreasing tendency, which converges in the later stage, but the final loss value is still high, indicating that there is still room for optimization. Our dynamic ArcFace loss function not only maintains the convergence of the original ArcFace loss but also pays more attention to hard samples, thus reducing the final loss value of the model and further proving that its self-learning weight can adjust the loss value in the network training process. The curve of the curricular loss accelerates the convergence of the model (the model converges after 30,000 iterations), but it does not reduce the loss value as much as the dynamic ArcFace loss. It can be inferred that the curricular loss has limited optimization effects on the model. In general, the experiments above prove the superiority of the dynamic ArcFace loss and prove the feasibility that the dynamic ArcFace loss can be used as the loss function of our model.
Table 3

Different loss functions

Loss functionRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
L-softmax loss44.9467.5238.22
A-softmax loss38.5763.2527.11
Cosine loss47.4671.9338.77
ArcFace loss56.5879.4747.91
Curricular loss45.7968.6038.19
Dynamic ArcFace loss65.9685.2657.10
Different loss functions

GC-BPCD module

To determine the effect the GC-BPCD module has on the basic branch, we performed several correlational experiments. We used the combination of the mask and the original image as the input of the basic branch and dynamic ArcFace loss as the loss function with the branch network and feature fusion excluded. Table 4 reveals that, among the six modules embedded in the basic branch, the results of the LCANet module and GC-BPCD module exceed those of other modules, while the GC-BPCD module offers superior performance with a rank-1 accuracy of 75.53%. Compared with several other channel attention mechanisms (except the GC-BPCD module), the attention module in LCANet improves the ability of the channel attention mechanism in the original SENet to capture the linear dependency cross-channels and can selectively emphasize the key information that is effective for the recognition task. The global attention mechanism in GCNet has shortcomings in global contextual modeling and channel attention modeling. Our GC-BPCD module not only improves GCNet’s ability to model the global contextual information in the feature map but also fully captures the channel dependency between any two channels.
Table 4

Attention modules

Attention moduleRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
SENet51.1475.5341.15
CBAM50.5375.9640.93
LCANet68.6089.0460.14
ECA-Net52.1175.9744.04
GCNet64.7485.9657.63
Proposed module75.5390.8767.82
Attention modules We further select three different images randomly from the training set and use Grad-CAM for visualization. Figure 8 shows that the response areas obtained by the attention modules in SENet, CBAM, and ECA-Net are relatively scattered, while large-scale response areas appear in the target tooth area and the background area (such as SENet and CBAM), and these modules fail to exclude the interference caused by the background area, thus reducing the rank-1 accuracy. The heat produced by the GCNet module is not focused on target areas either, owing to the deficiencies the model has on contextual modeling and channel attention modeling. The LCANet module has improved the attention mechanism in SENet and focuses more attention on the tooth area. Our GC-BPCD module inherits this tendency, aggregates the contextual information of the target area, pays more attention to the target area, and pays less attention to the background.
Figure 8

Grad-CAM visualizations

The first row contains three different images, and in the following rows, each row contains three heatmaps correspond to different images.

Grad-CAM visualizations The first row contains three different images, and in the following rows, each row contains three heatmaps correspond to different images. The architecture of our GC-BPCD module is quite clear and simple. It compensates for the deficiencies of the global attention mechanism in the original GCNet and strengthens its ability to capture channel dependencies. The experiments above prove that the GC-BPCD module is a lightweight but effective unit in recognition tasks.

Ablation studies in mask branch

In this section, we conduct more experiments to evaluate our mask branch. The basic branch is added since the mask branch cannot work as a single model, and the parameters in the basic branch are set to be consistent with the final version.

Backbone

Table 5 shows the results of different backbones implemented in the mask branch without feature fusion. Among these results, U-Net achieves better performance than ResNet, probably because U-Net is designed specifically for segmentation tasks and derives mask features more precisely. This finding inspired us to design our mask branch model. Otherwise, the model yields better performance after we replace BN with GN, demonstrating that the representation ability of the mask branch has been strengthened. Since the mask branch uses a mask (which only contains two values) as input, GN adjusts the local feature distribution and makes features extracted from masks more discriminative.
Table 5

Backbones

Backbone moduleRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
ResNet56.9271.8348.52
U-Net61.3074.1153.38
Proposed67.6277.5362.34
Proposed + GN69.8287.0161.50
Backbones

Feature fusion

As shown in Table 6, after feature fusion is aggregated, a huge leap occurs in the accuracy, which implies that the power of the mask branch is limited if added directly. When we use elementwise multiplication instead of elementwise addition, the rank-1 accuracy increases by 6.67%. The elementwise addition operation adds redundancy while multiplication emphasizes features selectively. Weight features tighten the relationship between basic features and mask features, aggregate them, and optimize features together. They adjust the training procedure automatically and can be seen as weights of basic features, similar to convolution filters in dynamic filter networks.
Table 6

Feature fusion

Feature fusionRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
No feature fusion69.8287.0161.50
Elementwise addition81.1493.5173.95
Elementwise multiplication87.8196.6782.92
Feature fusion

Loss function

The different choices of loss functions for the mask loss, including the L1 loss, softmax loss, and our dynamic ArcFace loss (improved loss), are given in Table 7, while the basic loss and weight loss are fixed as the dynamic ArcFace loss. As shown in Table 7, the accuracy of the softmax loss is lower than the accuracies of the other methods, while the L1 loss and dynamic ArcFace loss achieve comparable performance. Moreover, the type of loss function adopted in the mask loss affects the optimization of the basic branch. We visualize basic loss in Figure 9. The figure shows that the convergence trends for all the loss functions are the same, and the value of the L1 loss is lower. Additionally, the loss curve is smoother than that in Figure 7, indicating that the training procedure becomes easier.
Table 7

Loss functions for mask loss

Loss functionRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
Softmax83.1694.6578.06
Dynamic ArcFace loss86.0596.2381.32
L187.8196.6782.92
Figure 9

Loss curves of three different loss functions

Best viewed in color.

Loss functions for mask loss Loss curves of three different loss functions Best viewed in color.

Other classification models

We then evaluate several other classification or identification models on our dataset, and our model obtains the top accuracy. The results are shown in Table 8. First, we compare our model with classical classification models, including ResNet, ResNeXt, InceptionNet, DenseNet, and EfficientNet. Among these classical models, EfficientNet outperforms other models, and the performance of InceptionNet and DenseNet is not satisfactory. The reason is that InceptionNet and DenseNet are deeper than the others. DenseNet reuses parameters by concatenating channels in each dense block but loses information when 1 × 1 convolutions are used, and EfficientNet is embedded with the attention model in SENet, thus concentrating more on tooth areas.
Table 8

Other models

ModelRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
ResNet54.2177.7242.97
ResNeXt58.6881.1450.27
InceptionNet48.777539.03
DenseNet44.5671.0534.32
EfficientNet69.2187.2061.35
Ajaz and Kathirvelu17.5124.398.73
Oktay26.7547.3718.00
DentNet39.4765.0928.17
LCANet78.8692.8172.67
Proposed87.8196.6782.92
Other models Moreover, compared with the other four identification networks, our model performs well. Ajaz and Kathirvelu and Oktay are two classical methods applied on small datasets consisting of 200 images and have poor effects. DenseNet is a deep learning model composed of a backbone network similar to ResNet without dense connections, and its performance is not as good as those of LCANet and our network. LCANet aggregates self-learning weight features and attention mechanisms and has a better effect, but it does not make full use of dental morphological information.

Cross-validation

To further evaluate our model, we perform 5-fold cross-validation, and the results are shown in Table 9. The final average rank-1 accuracy is 88.62%, the highest rank-1 accuracy is 89.91%, and the lowest is 87.74%. All these data indicate that our model performs well and has generalization ability.
Table 9

Five-fold cross-validation

NumberRank-1 accuracyRank-10 accuracyTAR(@FAR = 10-4)
189.9196.5783.90
288.6896.3283.88
387.7495.6783.43
488.4396.0886.27
588.3496.1783.23
Average88.6296.1684.14
Five-fold cross-validation

Experimental conclusion

In the section “ablation studies in basic branch,” we perform ablation studies in the basic branch. In the section “masks,” we explore the role of masks. With other modified units abandoned, when masks are added, rank-1 accuracy increases from 34.47 to 38.94 by four points and is still low. We substitute softmax loss with other loss functions in the section “dynamic ArcFace loss (improved loss).” The dynamic ArcFace loss exceeds ArcFace loss by approximately nine points, which proves that it fully exploits the potential of the added masks. Next, we compare several channel attention modules. The proposed model outperforms the LCA module by approximately seven points, possibly because the proposed model fully captures the channel dependency between any two channels. A detailed analysis is provided in the section “GC-BPCD module.” The rank-1 accuracy reaches 75.53 when only the basic branch is employed. Ablation studies in the mask branch are performed in section “ablation studies in mask branch.” First, in section “backbone,” we compare backbone models (mask extractor in Figure 1). The accuracy of the proposed backbone model is 69.82, which outperforms other backbone models. There is no feature fusion (green part in Figure 1) at that stage, and the mask branch extracts features of masks and makes predictions in categories. Owing to the large number of classes, the model is not very accurate in its predictions, so the accuracy decreases to 69.82 from its value of 75.53 in section “ablation studies in basic branch.” Next, in section “feature fusion,” we perform different methods of feature fusion. Elementwise multiplication bridges the gap between images and masks and increases the accuracy by a large amount to 87.81. We compare different loss functions for mask loss and achieve a higher accuracy of L1 loss. In section “other classification models,” we evaluate other classification or identification models. In section “cross-validation,” we perform 5-fold cross-validation and obtain an average rank-1 accuracy of 88.62. The aforementioned ablation studies prove the effectiveness of every single part of our model. The usage of teeth masks is the highlight of the proposed model, but a simple combination just trails the desired accuracy. Directly concatenating masks with images (section “masks”) or barely adding lateral connections (section “backbone”) does not fully utilize masks, which limits the overall performance. Every part of the proposed model is devised to address the fusion problem. There also exist limitations. Teeth have an important role in the proposed model. The proposed model rests on masks, and extra time is needed to draw masks of teeth. In subsequent studies, we will perform unsupervised methods on teeth or other structures.

Conclusion

Formal works on human identification with panoramic dental images aim for classification but neglect the effect of dental morphology. Based on this limitation, we propose a novel fine-grained bilateral-branch network. Our key idea is to promote the feature fusion of images and masks, thus to help the model focus on meaningful tooth contours. A novel attention module (the GC-BPCD module) is proposed to localize discriminative parts, and an improved loss (dynamic ArcFace loss) is proposed to address easy samples and hard samples. Ablation studies prove that the proposed model outperforms other identification models devised specifically for the dataset and other general classification models. Experiments elaborate that the proposed method is robust and easy to converge.
  6 in total

Review 1.  The scope of forensic radiology.

Authors:  B G Brogdon
Journal:  Clin Lab Med       Date:  1998-06       Impact factor: 1.935

2.  Transfer Learning Based Automatic Human Identification using Dental Traits- An Aid to Forensic Odontology.

Authors:  Sathya B; Neelaveni R
Journal:  J Forensic Leg Med       Date:  2020-09-29       Impact factor: 1.614

3.  Automatic human identification from panoramic dental radiographs using the convolutional neural network.

Authors:  Fei Fan; Wenchi Ke; Wei Wu; Xuemei Tian; Tu Lyu; Yuanyuan Liu; Peixi Liao; Xinhua Dai; Hu Chen; Zhenhua Deng
Journal:  Forensic Sci Int       Date:  2020-07-15       Impact factor: 2.395

4.  Co-Learning Feature Fusion Maps from PET-CT Images of Lung Cancer.

Authors:  Ashnil Kumar; Michael Fulham; Dagan Feng; Jinman Kim
Journal:  IEEE Trans Med Imaging       Date:  2019-06-17       Impact factor: 10.048

5.  LCANet: Learnable Connected Attention Network for Human Identification Using Dental Images.

Authors:  Yancun Lai; Fei Fan; Qingsong Wu; Wenchi Ke; Peixi Liao; Zhenhua Deng; Hu Chen; Yi Zhang
Journal:  IEEE Trans Med Imaging       Date:  2021-03-02       Impact factor: 10.048

6.  ArcFace: Additive Angular Margin Loss for Deep Face Recognition.

Authors:  Jiankang Deng; Jia Guo; Jing Yang; Niannan Xue; Irene Kotsia; Stefanos Zafeiriou
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2022-09-14       Impact factor: 9.322

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.