| Literature DB >> 32599883 |
Kh Tohidul Islam1,2, Ram Gopal Raj1, Syed Mohammed Shamsul Islam3,4, Sudanthi Wijewickrema2, Md Sazzad Hossain5, Tayla Razmovski2, Stephen O'Leary2.
Abstract
Automatic vehicle license plate recognition is an essential part of intelligent vehicle access control and monitoring systems. With the increasing number of vehicles, it is important that an effective real-time system for automated license plate recognition is developed. Computer vision techniques are typically used for this task. However, it remains a challenging problem, as both high accuracy and low processing time are required in such a system. Here, we propose a method for license plate recognition that seeks to find a balance between these two requirements. The proposed method consists of two stages: detection and recognition. In the detection stage, the image is processed so that a region of interest is identified. In the recognition stage, features are extracted from the region of interest using the histogram of oriented gradients method. These features are then used to train an artificial neural network to identify characters in the license plate. Experimental results show that the proposed method achieves a high level of accuracy as well as low processing time when compared to existing methods, indicating that it is suitable for real-time applications.Entities:
Keywords: artificial neural networks; automatic license plate recognition; histogram of oriented gradients; intelligent vehicle access
Year: 2020 PMID: 32599883 PMCID: PMC7349508 DOI: 10.3390/s20123578
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Outline of the proposed system for automatic license plate recognition.
Image acquisition properties.
| Name | Description |
|---|---|
| Image acquisition device name | Canon® Power Shot SX530 HS |
| Zooming capabilities | 50× Optical zoom |
| Camera zooming position | 5× Optical zoom |
| Weather | Daylight, rainy, sunny, cloudy |
| Capturing period | Day and Night |
| Background | Complex; not fixed |
| Horizontal field-of-view | Approximately 75° |
| Image dimension |
|
| Vehicle speed limit | 20 km/h; 5.56 m/s |
| Capturing distance | 15 meter |
Figure 2The initial extracted region of interest, defined by the green bounding box.
Figure 3The architecture of YOLO V2 network.
Figure 4Detection result: from left to right, input image, ground truth image (red bounding box), and output image with detected license plate ROI (green bounding box).
Figure 5Alphanumeric character segmentation: (A) before gray level histogram equalization; (B) after gray level histogram equalization; (C) median filtered and noise removed image; (D) binary image; (E) character region segmentation; and (F) extracted and resized characters.
Figure 6HOG feature vector visualization with different cell sizes.
Figure 7Proposed recognition process along with the artificial neural network architecture.
Figure 8Synthetic samples for training purposes.
Training performance with respect to hidden neuron size. The best performance is highlighted in bold.
| Hidden Neurons | Repetition Number | Iterations | Time | Performance | Gradient | Error (%) |
|---|---|---|---|---|---|---|
| 10 | 1 | 160 | 0:00:45 | 1.47 × 10 | 9.03 × 10 | 1.50 × 10 |
| 2 | 179 | 0:00:51 | 8.10 × 10 | 9.09 × 10 | 6.11 × 10 | |
| 3 | 273 | 0:01:17 | 6.00 × 10 | 9.62 × 10 | 8.06 × 10 | |
| 4 | 189 | 0:00:53 | 2.02 × 10 | 1.34 × 10 | 1.86 × 10 | |
| 5 | 214 | 0:01:00 | 1.34 × 10 | 6.66 × 10 | 1.19 × 10 | |
| 20 | 1 | 167 | 0:01:07 | 2.00 × 10 | 1.04 × 10 | 2.50 × 10 |
| 2 | 186 | 0:01:18 | 4.86 × 10 | 1.97 × 10 | 2.50 × 10 | |
| 3 | 140 | 0:00:59 | 1.51 × 10 | 6.95 × 10 | 3.33 × 10 | |
| 4 | 165 | 0:01:10 | 6.12 × 10 | 2.72 × 10 | 2.22 × 10 | |
| 5 | 137 | 0:00:58 | 1.04 × 10 | 5.59 × 10 | 3.61 × 10 | |
|
| 1 | 124 | 0:01:26 | 3.65 × 10 | 2.35 × 10 | 2.78 × 10 |
| 2 | 124 | 0:01:25 | 2.78 × 10 | 2.12 × 10 | 1.67 × 10 | |
|
|
|
|
|
|
| |
| 4 | 142 | 0:02:03 | 7.50 × 10 | 3.67 × 10 | 1.67 × 10 | |
| 5 | 105 | 0:01:18 | 1.22 × 10 | 1.81 × 10 | 3.33 × 10 |
The number of characters extracted from license plate images in each class.
| Characters | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | A |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Quantity | 29 | 37 | 36 | 40 | 40 | 41 | 39 | 51 | 35 | 33 | 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| Quantity | 16 | 7 | 7 | 10 | 5 | 7 | 7 | 13 | 12 | 12 | 7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| Quantity | 10 | 15 | 9 | 12 | 9 | 13 | 3 | 15 | 73 | 6 | 9 |
Figure 9Real-time experimental results for license plate character recognition.
Comparison of performance for different combinations of feature extraction and classification methods. The best performance per metric is highlighted in bold.
| Method | Processing Time (s) | Accuracy (%) | ||||
|---|---|---|---|---|---|---|
| Plate | Character | Feature | Classification | Total | ||
| BoF+SAE | 0.15 | 0.25 | 0.21 | 0.015 | 0.625 | 95.73 |
| BoF+KNN | 0.15 | 0.25 | 0.21 | 0.024 | 0.634 | 88.25 |
| BoF+SVM | 0.15 | 0.25 | 0.21 | 0.020 | 0.630 | 89.78 |
| BoF+ANN | 0.15 | 0.25 | 0.21 | 0.021 | 0.631 | 98.33 |
| SIFT+SAE | 0.15 | 0.25 | 0.28 | 0.019 | 0.699 | 93.75 |
| SIFT+KNN | 0.15 | 0.25 | 0.28 | 0.030 | 0.710 | 87.38 |
| SIFT+SVM | 0.15 | 0.25 | 0.28 | 0.027 | 0.707 | 88.94 |
| SIFT+ANN | 0.15 | 0.25 | 0.28 | 0.026 | 0.706 | 96.18 |
| HOG+SAE | 0.15 | 0.25 | 0.27 | 0.018 | 0.688 | 94.30 |
| HOG+KNN | 0.15 | 0.25 | 0.27 | 0.028 | 0.698 | 97.60 |
| HOG+SVM | 0.15 | 0.25 | 0.27 | 0.025 | 0.695 | 98.90 |
|
|
|
|
|
|
|
|
The classification performance for the proposed method compared to similar existing methods. The best performance per metric is highlighted in bold.
| Method | Feature Extraction Method | Classifier | Total Time (s) | Accuracy (%) |
|---|---|---|---|---|
| Jin et al. [ | Hand-Crafted | Fuzzy | 0.432 | 92.00 |
| Arafat et al. [ | OCR | OCR | 0.681 | 97.86 |
| Samma et al. [ | Haar-like | FSVM | 0.649 | 98.36 |
| Tabrizi et al. [ | KNN+SVM | KNN+SVM | 0.721 | 97.03 |
| Niu et al. [ | HOG | SVM | 0.645 | 96.60 |
| Li et al. [ | CNN | CNN | 0.825 | 99.20 |
| Thakur et al. [ | GA | ANN | 0.532 | 97.00 |
| Cheng et al. [ | SCDCS-LS | RWNN | 0.659 | 99.54 |
| Lee et al. [ | AlexNet | AlexNet | 0.983 | 99.58 |
|
|
|
|
|
|
Performance comparison on the Medialab LPR database. The stages that were not addressed in the original papers are denoted by “—”. The best performance per metric is highlighted in bold.
| Method | Accuracy (%) | ||
|---|---|---|---|
| Detection | Segmentation | Classification | |
| Jin et al. [ | 95.73 | 98.87 | 91.25 |
| Arafat et al. [ | 98.30 | 99.30 | 96.57 |
| Samma et al. [ | 96.25 | — | 98.05 |
| Tabrizi et al. [ | 96.98 | 96.85 | 96.54 |
| Niu et al. [ | 98.45 | — | 96.38 |
| Li et al. [ | — | — | 98.52 |
| Thakur et al. [ | 97.85 | 98.37 | 97.35 |
| Cheng et al. [ | — | — | 99.38 |
| Lee et al. [ | — | — | 97.38 |
|
|
|
|
|
Performance comparison on the UFPR-ALPR database. The stages that were not addressed in the original papers are denoted by “—”. The best performance per metric is highlighted in bold.
| Method | Accuracy (%) | ||
|---|---|---|---|
| Detection | Segmentation | Classification | |
| Jin et al. [ | 85.48 | 91.75 | 85.35 |
| Arafat et al. [ | 85.45 | 93.45 | 90.37 |
| Samma et al. [ | 80.35 | — | 91.70 |
| Tabrizi et al. [ | 84.45 | 90.50 | 92.86 |
| Niu et al. [ | 85.80 | — | 89.32 |
| Li et al. [ | — | — | 92.71 |
| Thakur et al. [ | 82.35 | 91.22 | 90.85 |
| Cheng et al. [ | — | — | 92.50 |
| Lee et al. [ | — | — | 92.75 |
|
|
|
|
|