| Literature DB >> 27999301 |
Husan Vokhidov1, Hyung Gil Hong2, Jin Kyu Kang3, Toan Minh Hoang4, Kang Ryoung Park5.
Abstract
Automobile driver information as displayed on marked road signs indicates the state of the road, traffic conditions, proximity to schools, etc. These signs are important to insure the safety of the driver and pedestrians. They are also important input to the automated advanced driver assistance system (ADAS), installed in many automobiles. Over time, the arrow-road markings may be eroded or otherwise damaged by automobile contact, making it difficult for the driver to correctly identify the marking. Failure to properly identify an arrow-road marker creates a dangerous situation that may result in traffic accidents or pedestrian injury. Very little research exists that studies the problem of automated identification of damaged arrow-road marking painted on the road. In this study, we propose a method that uses a convolutional neural network (CNN) to recognize six types of arrow-road markings, possibly damaged, by visible light camera sensor. Experimental results with six databases of Road marking dataset, KITTI dataset, Málaga dataset 2009, Málaga urban dataset, Naver street view dataset, and Road/Lane detection evaluation 2013 dataset, show that our method outperforms conventional methods.Entities:
Keywords: advanced driver assistance system (ADAS); arrow-road marking recognition; convolutional neural network; damaged arrow-road marking; visible light camera sensor
Year: 2016 PMID: 27999301 PMCID: PMC5191139 DOI: 10.3390/s16122160
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Comparisons of previous and proposed methods.
| Category | Methods | Advantages | Disadvantages | Performance | Year | Ref. |
|---|---|---|---|---|---|---|
| Non-learning-based | Geometric parameter optimization | Fast processing speed | The number of thresholds to be set is large | True positive rate (TPR) of 90% and 78% for crosswalks and arrows, respectively | 2011 | [ |
| HOG features and template matching | A good ability to cope with the cases of limited occlusions and the variations of lighting condition | In bright conditions, system is sensitive to shadows High FPR for simple but important signs such as forward arrows | TPR of 90.1% and false positive rate (FPR) 0.9% | 2012 | [ | |
| Template matching | Fast processing speed | Damaged road marking can cause misclassification, and the classification accuracy is affected by the illumination variation | Detection rate of 95.8% and 84% on the highway and city roads, respectively | 2012 | [ | |
| Learning-based | HOG features and total error rate (TER)-based classifier | Fast computing time compared to SVM-based method | Damaged or shadowed markings increase FPR | Overall classification accuracy of 99.2% | 2015 | [ |
| HOG features and SVM | Showing high accuracy with the trained datasets | Recognition accuracy can be affected by damaged or shadowed markings | Quantitative accuracies were not reported | 2015 | [ | |
| F-measure of 0.91 | 2015 | [ | ||||
| Average accuracy of 91.7% | 2014 | [ | ||||
| Fourier descriptor and KNN classifier | Robust to noises on road marking | Sensitive to occlusion, dirty markings or poor visibility | Average error of 6% | 2004 | [ | |
| Artificial Neural Network | Higher accuracy with trained datasets compared to non-learning-based method | Performance of testing data can be affected by trained dataset | Quantitative accuracies were not reported | 1994 | [ | |
| The average accuracy of white markings is about 71.5%, and for orange markings it was about 46% | 2014 | [ | ||||
| Accuracy of 85% for arrows | 2012 | [ | ||||
| BING, PCA network, and SVM classifier | The area of road marking can be detected by BING method without lane detection | Performance of testing data can be affected by trained dataset | Accuracy of 96.8% | 2015 | [ | |
| Arrow-road markings in various environments including damaged ones can be correctly recognized independent of the kinds of datasets by intensive training of CNN | Time consuming procedure for training is required for CNN | Average accuracy and F_score are 99.88% and 99.94%, respectively |
Figure 1Overall flowchart of proposed method.
The CNN architecture used in our research.
| Layer Type | Number of Filters | Size of Feature Map | Size of Kernel | Number of Stride |
|---|---|---|---|---|
| Image input layer | 265 (height) × 137 (width) × 1 (channel) | |||
| 1st convolutional layer | 180 | 131 × 67 × 180 | [5 5] | [2 2] |
| ReLU layer | ||||
| CCN layer | ||||
| Max pooling layer | 180 | 65 × 33 × 180 | [3 3] | [2 2] |
| 2nd convolutional layer | 250 | 31 × 15 × 250 | [5 5] | [2 2] |
| ReLU layer | ||||
| CCN layer | ||||
| Max pooling layer | 250 | 15 × 7 × 250 | [3 3] | [2 2] |
| 3rd convolutional layer | 250 | 7 × 3 × 250 | [3 3] | [2 2] |
| ReLU layer | ||||
| CCN layer | ||||
| Max pooling layer | 250 | 3 × 1 × 250 | [3 3] | [2 2] |
| 1st fully connected layer | 1920 | |||
| ReLu layer | ||||
| 2nd fully connected layer | 1024 | |||
| ReLu layer | ||||
| 3rd fully connected layer | 512 | |||
| ReLu layer | ||||
| Dropout layer | ||||
| 4th fully connected layer | 6 | |||
| Softmax layer | ||||
| Classification layer (output layer) |
Figure 2Convolutional neural network (CNN) architecture.
Figure 3Example images from Road marking dataset.
Figure 4Example images from Karlsruhe institute of technology and Toyota technological institute at Chicago (KITTI) dataset.
Figure 5Example images from Málaga dataset 2009.
Figure 6Example images from Málaga urban dataset.
Figure 7Example images from Naver street view dataset.
Figure 8Example images from Road/Lane detection evaluation 2013 dataset.
Figure 9Obtaining the inverse perspective mapping (IPM) image and arrow markings: (a) original image; (b) IPM transformation; and (c) the obtained arrow markings.
Figure 10Examples of arrow markings after size normalization and bi-linear interpolation.
Number of data for our experiments.
| FA | FLA | FLRA | FRA | LA | RA | Total | |
|---|---|---|---|---|---|---|---|
| Number of data | 32,686 | 17,885 | 22,344 | 17,885 | 36,766 | 36,243 | 163,809 |
Figure 11Classification accuracies of training data over 13-fold cross validation; “classifications” means “classification accuracy of training data”: (a) Training 1–3; (b) Training 4–6; (c) Training 7–9; (d) Training 10–13.
Figure 12The obtained filters from the 1st convolution layer through training: (a–m) the filters from the 1st–13th trainings among 13-fold cross validation are presented, respectively.
The summated confusion matrix of tests 1–13.
| Total of Testing 1–13 | Recognized Arrows | ||||||
|---|---|---|---|---|---|---|---|
| FA | FLA | FLRA | FRA | LA | RA | ||
| Actual arrows | FA | 31,447 | 1 | 0 | 1 | 0 | 2 |
| FLA | 0 | 18,101 | 0 | 1 | 0 | 0 | |
| FLRA | 21 | 0 | 22,254 | 0 | 0 | 0 | |
| FRA | 0 | 0 | 0 | 18,824 | 24 | 0 | |
| LA | 1 | 0 | 0 | 0 | 33,334 | 210 | |
| RA | 0 | 0 | 0 | 0 | 0 | 31,779 | |
Accuracies of arrow marking recognition by our method (unit: %).
| # of Testing | FA | FLA | FLRA | FRA | LA | RA | |
|---|---|---|---|---|---|---|---|
| Testing 1 | ACC | 100 | 100 | 100 | 100 | 99.45 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
| Testing 2 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
| Testing 3 | ACC | 99.91 | 100 | 100 | 100 | 99.05 | 100 |
| F_score | 99.96 | 100 | 100 | 100 | 99.52 | 100 | |
| Testing 4 | ACC | 100 | 100 | 100 | 100 | 99.02 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.51 | 100 | |
| Testing 5 | ACC | 100 | 100 | 100 | 100 | 99.12 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.56 | 100 | |
| Testing 6 | ACC | 100 | 100 | 100 | 100 | 99.41 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.70 | 100 | |
| Testing 7 | ACC | 99.96 | 100 | 100 | 100 | 99.34 | 100 |
| F_score | 99.98 | 100 | 100 | 100 | 99.67 | 100 | |
| Testing 8 | ACC | 100 | 100 | 100 | 99.11 | 100 | 100 |
| F_score | 100 | 100 | 100 | 99.55 | 100 | 100 | |
| Testing 9 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
| Testing 10 | ACC | 99.96 | 100 | 100 | 100 | 99.48 | 100 |
| F_score | 99.98 | 100 | 100 | 100 | 99.74 | 100 | |
| Testing 11 | ACC | 100 | 100 | 99.22 | 100 | 100 | 100 |
| F_score | 100 | 100 | 99.61 | 100 | 100 | 100 | |
| Testing 12 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
| F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
| Testing 13 | ACC | 100 | 99.93 | 100 | 100 | 98.99 | 100 |
| F_score | 100 | 99.96 | 100 | 100 | 99.49 | 100 | |
| Average ACC | 99.99 | 99.99 | 99.94 | 99.93 | 99.40 | 100 | |
| Average F_score | 99.99 | 99.997 | 99.97 | 99.97 | 99.70 | 100 | |
Comparisons of accuracies of recognition of arrow marking by our method with previous method (unit: %).
| Our Method | Previous Method [ | |
|---|---|---|
| Average ACC | 99.88 | 92.8 |
| Average F_score | 99.94 | 93.9 |
Figure 13Examples of correct recognition cases: (a) forward arrow (FA); (b–d) forward-left arrow (FLA); (e) forward-right arrow (FRA); (f) left arrow (LA); (g) right arrow (RA).
Figure 14Examples of incorrect recognition cases: (a) FA is incorrectly recognized into FLA; and (b,c) LA is incorrectly recognized into RA.
The summated confusion matrix by revised CNN.
| Total of Testing 1–13 | Recognized Arrows | ||||||
|---|---|---|---|---|---|---|---|
| FA | FLA | FLRA | FRA | LA | RA | ||
| Actual arrows | FA | 31,451 | 0 | 0 | 0 | 0 | 0 |
| FLA | 0 | 18,102 | 0 | 0 | 0 | 0 | |
| FLRA | 0 | 0 | 22,275 | 0 | 0 | 0 | |
| FRA | 0 | 0 | 0 | 18,848 | 0 | 0 | |
| LA | 0 | 0 | 0 | 0 | 33,532 | 13 | |
| RA | 0 | 0 | 0 | 0 | 0 | 31,779 | |
Accuracies of recognition of arrow marking by revised CNN (unit: %).
| Our Method | |
|---|---|
| Average ACC | 99.99 |
| Average F_score | 99.99 |