| Literature DB >> 31652552 |
Ahmed Gomaa1,2,3, Moataz M Abdelwahab4, Mohammed Abo-Zahhad5,6, Tsubasa Minematsu7, Rin-Ichiro Taniguchi8.
Abstract
Automatic vehicle detection and counting are considered vital in improving traffic control and management. This work presents an effective algorithm for vehicle detection and counting in complex traffic scenes by combining both convolution neural network (CNN) and the optical flow feature tracking-based methods. In this algorithm, both the detection and tracking procedures have been linked together to get robust feature points that are updated regularly every fixed number of frames. The proposed algorithm detects moving vehicles based on a background subtraction method using CNN. Then, the vehicle's robust features are refined and clustered by motion feature points analysis using a combined technique between KLT tracker and K-means clustering. Finally, an efficient strategy is presented using the detected and tracked points information to assign each vehicle label with its corresponding one in the vehicle's trajectories and truly counted it. The proposed method is evaluated on videos representing challenging environments, and the experimental results showed an average detection and counting precision of 96.3% and 96.8%, respectively, which outperforms other existing approaches.Entities:
Keywords: background subtraction; deep convolutional neural network; intelligent transportation system; vehicle counting; vehicle dtection
Year: 2019 PMID: 31652552 PMCID: PMC6832389 DOI: 10.3390/s19204588
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Evaluation summary with the state of art.
| Compared Algorithm | Advantages | Disadvantages |
|---|---|---|
| Corola [ | Low computational complexity. | Only detection method. |
| Yang et al. [ | Low computational complexity. | Declined counting precision accuracy. |
| Quesada et al. [ | Low computational complexity. | |
| Mohamed [ | Faster counting strategy. | |
| Proposed Method | Low computational complexity. |
Figure 1Block diagram of the proposed approach for a specific frameset S.
The structure of the adopted convolutional neural network, where K is the size of the kernel, S is the stride [16].
| Layer Type | Parameters |
|---|---|
| Input | |
| Convolution | 6 filters, K: |
| Activate function | ReLU |
| Maxpooling | K: |
| Convolution | 16 filters, K: |
| Activate function | ReLU |
| Maxpooling | K: |
| Fully connected | 120 hidden units |
| Sigmoid | 2 classes ( Foreground/Background ) |
Figure 2Regions of detection.
Figure 3Feature points in the detected region of interest.
Figure 4Clustering the foreground vehicles.
Figure 5Connecting vehicles trajectories.
Figure 6Variable bounding box.
Figure 7Varying bounding box coordinates.
Challenge environments information of the sequences used in the performance evaluation.
| Dataset | GRAM Dataset | CDnet2014 | ATON Testbed | ||||
|---|---|---|---|---|---|---|---|
| Sequence | M-30 | M-30-HD | Highway | Intermittenpan | Streetcorneratnight | Tramstation | Highway II |
| Challenging | Sunny day, | High resolution | Sunny day, | Sunny day, | Light changes, | Night scene, | Crowed scene. |
Vehicle counting accuracy for first experiment.
| Compared Algorithm | GRAM Dataset | ATON Testbed | ||||
|---|---|---|---|---|---|---|
| M-30 | M-30-HD | Highway II | ||||
| Miss Detection | Precision | Miss Detection | Precision | Miss Detection | Precision | |
| Yang et al. [ | 6 | 92.20 | 5 | 88.10 | 3 | 92.31 |
| Quesada et al. [ | 2 | 97.41 | 3 | 92.86 | N/A | N/A |
| Mohamed [ | 1 | 98.70 | 0 | 100 | 2 | 95.65 |
| Proposed Method | 0 | 100 | 0 | 100 | 1 | 97.9 |
Vehicle detection results comparison on CDnet2014 sequences.
| Method | Corola [ | Yang et al. [ | Proposed Method | |
|---|---|---|---|---|
| Sequence Videos | ||||
| Highway | Precission % | 95.1 | 91.3 | 98.2 |
| Recall % | 85.4 | 92.1 | 99.2 | |
| Intermittenpan | Precission % | 56.6 | 90.2 | 98.7 |
| Recall % | 58.5 | 98.8 | 97.5 | |
| Streetcorneratnight | Precission % | 82.4 | 89.1 | 95.8 |
| Recall % | 86.5 | 97.2 | 97.1 | |
| Tramstation | Precission % | 44.7 | 87.1 | 92.3 |
| Recall % | 91.0 | 97.1 | 97.1 | |
| Average accuracy | Precission % | 69.7 | 89.4 | 96.3 |
| Recall % | 80.4 | 96.3 | 97.7 | |
Vehicle counting accuracy for CDnet2014 sequences.
| Compared Algorithm | Highway Precision | Intermittenpan Precision | Streetcorneratnight Precision | TramStation Precision |
|---|---|---|---|---|
| Yang et al. [ | 93.3 | 93.3 | 90.4 | 84.6 |
| Mohamed [ | 92.3 | N/A | N/A | N/A |
| Proposed Method | 100 | 93.3 | 95.2 | 91.6 |
Figure 8Sample results on GRAM and CDnet2014 dataset. Top row: Cloudy and crowded. Second and fourth row: Waving trees. Third and Bottom row: Night scene with changed light.