| Literature DB >> 35725753 |
Weidong Li1,2, Shuai Wang3,4, Inam Ullah1,2, Xuehai Zhang1,2, Jinlong Duan1,2.
Abstract
Factories swiftly and precisely grasp the real-time data of the production instrumentation, which is the foundation for the development and progress of industrial intelligence in industrial production. Weather, light, angle, and other unknown circumstances, on the other hand, impair the image quality of meter dials in natural environments, resulting in poor dial image quality. The remote meter reading system has trouble recognizing dial pictures in extreme settings, challenging it to meet industrial production demands. This paper provides multiple attention and encoder-decoder-based gas meter recognition networks (MAEDR) for this problem. First, from the acquired dial photos, the dial images with extreme conditions such as overexposure, artifacts, blurring, incomplete display of characters, and occlusion are chosen to generate the gas meter dataset. Then, a new character recognition network is proposed utilizing multiple attention and an encoder-decoder structure. Convolutional neural networks (CNN) extract visual features from dial images, encode visual features employing multi-head self-attention and position information, and facilitate feature alignment using the connectionist temporal classification (CTC) method. A novel two-step attention decoder is presented to improve the accuracy of recognition results. convolutional block attention module (CBAM) reweights the visual features from the CNN and the semantic features computed by the encoder to improve model performance; long short-term memory attention (LSTM attention) focuses on the relationship between feature sequences. According to experimental data, our system can effectively and efficiently identify industrial gas meter picture digits with 91.1% identification accuracy, faster inference speed, and higher accuracy than standard algorithms. The accuracy and practicality of the recognition can fulfill the needs of instrument data detection and recognition in industrial production, and it has a wide range of applications.Entities:
Mesh:
Year: 2022 PMID: 35725753 PMCID: PMC9209494 DOI: 10.1038/s41598-022-14434-0
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1Model of our proposed recognition system.
Figure 2K-means clustering results for the gas meter dataset.
Figure 3Architecture diagram of our two-step attention decoder.
Figure 4Performance comparison of MAEDR model (Ours-1) and MADER + CTC model (Ours-2).
Recognition performances of our method compared with other baselines on the gas meter dataset.
| Model | Accuracy (%) | FPS | Parameter number |
|---|---|---|---|
| CRNN | 87.5 | 142 | |
| SCATTER | 86.2 | 41 | 119375120 |
| SRN | 87.4 | 71 | 57199301 |
| HGA-STR | 83.4 | 17 | 33916725 |
| MAEDR (Ours-1) | 86.0 | ||
| MAEDR + CTC (Ours-2) |
In each column, the best performing result is shown in bold, and the second best result is shown in underline.
Recognition performances of our method compared with CRNN on the gas meter dataset.
Recognition performances of our method compared with SOTA on public datasets.
| Model | Test datasets & accuracy(%) | ||
|---|---|---|---|
| IIIT5K | SVT | IC13 | |
| SCATTER | 93.7 | 92.7 | 93.9 |
| CRNN | 82.9 | 81.6 | 89.2 |
| SRN | 94.8 | 91.5 | 95.5 |
| RobustScanner | 95.3 | 88.1 | 94.8 |
| ABINet-LV | 96.2 | 93.5 | |
| PRENatal2D | 95.6 | 94.0 | 96.4 |
| MAEDR (Ours-1) | 95.1 | 95.2 | |
| MAEDR + CTC (Ours-2) | |||
In each column, the best performing result is shown in bold, and the second best result is shown in underline.
Comparison of results of ablation experiments.
| Methods | Test datasets & accuracy(%) | |||
|---|---|---|---|---|
| Our | IIIT5K | SVT | IC13 | |
| MAEDR+CTC | ||||
| Without CTC | 86.0 | 95.1 | 94.1 | 95.2 |
| Without CBAM | 85.4 | 95.6 | 93.2 | 95.7 |
| Without CTC & CBAM | 84.2 | 94.3 | 92.5 | 94.1 |
In each column, the best performing result is shown in bold.