| Literature DB >> 35479608 |
Yanfeng Wang1, Yinan Chen2, Runmin Liu3,4.
Abstract
With the deepening of deep learning research, progress has been made in artificial intelligence. In the process of aircraft classification, the precision rate of aircraft picture recognition based on traditional methods is low due to various types of aircraft, large similarities between different models, and serious texture interference. In this article, the hybrid attention network model (BA-CNN) to implement an aircraft recognition algorithm is proposed to solve the above problems. Using two-channel ResNet-34 as a characteristic extraction function, the depth of network is increased to improve fine-grained characteristic extraction capability without increasing the output characteristic dimension. In the network to introduce a hybrid attention mechanism, respectively, between the residual units of two ResNet-34 channels, channel attention and spatial attention modules are added, more abundant mixed characteristics of attention are obtained, space and characteristics of the local characteristics of the channel response are focused, the characteristics of redundancy are reduced, and the fine-grained characteristics of learning ability are further enhanced. Trained and tested on FGVC-aircraft, a public fine-grained pictures dataset, the recognition precision rate of the BA-CNN networks model reached 89.2%. It can be seen from the experimental results, the recognition precision rate of the original model is improved effectively by using this method, and the recognition precision rate is higher than most of the existing mainstream aircraft recognition ways.Entities:
Mesh:
Year: 2022 PMID: 35479608 PMCID: PMC9038415 DOI: 10.1155/2022/4189500
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Sample labeling diagram.
Figure 2Selected aircraft sample pictures.
Figure 3Aircraft pictures before and after spatial domain enhancement.
Figure 4Hybrid attention network model architecture.
Figure 5Schematic diagram of characteristic fusion.
Figure 6Structure of residual unit.
Figure 7Channel attention module.
Figure 8Spatial attention module.
Figure 9Residual attention structure of BA-CNN network model.
Experimental analysis of the ablation of the method in this article on the FGVC-Aircraft dataset.
| Approach | Backbone | Accuracy (%) |
|---|---|---|
| BA-CNN (resnet × 2) | ResNet-34 × 2 | 85.0 |
| BA-CNN (channel attention) | ResNet-34 × 2 + channel attention | 86.2 |
| BA-CNN (spatial attention) | ResNet-34 × 2 + spatial attention | 86.5 |
| BA-CNN (channel and spatial attention) | ResNet-34 × 2 + channel and spatial attention | 89.2 |
Comparison of the recognition precision rate of different weakly supervised algorithms.
| Approach | Backbone | Precision rate (%) |
|---|---|---|
| Two-level attention | VGG19 | 77.9 |
| NAC | VGG19 | 81.01 |
| B-CNN | VGG-M + vgg-d | 84.1 |
| ST-CNN | Inception-v2 × 3 | 84.1 |
| DVAN | VGG-19 × 3 | 79.0 |
| RA-CNN | VGG-19 × 3 | 85.3 |
| MA-CNN | VGG-19 × 3 | 86.5 |
| MAMC | ResNet-101 | 86.5 |
| BA-CNN | ResNet-34 × 2 | 89.2 |