| Literature DB >> 36236780 |
Abdullah Al Mamun1, Em Poh Ping1, Jakir Hossen1, Anik Tahabilder2, Busrat Jahan3.
Abstract
Lane marking recognition is one of the most crucial features for automotive vehicles as it is one of the most fundamental requirements of all the autonomy features of Advanced Driver Assistance Systems (ADAS). Researchers have recently made promising improvements in the application of Lane Marking Detection (LMD). This research article has taken the initiative to review lane marking detection, mainly using deep learning techniques. This paper initially discusses the introduction of lane marking detection approaches using deep neural networks and conventional techniques. Lane marking detection frameworks can be categorized into single-stage and two-stage architectures. This paper elaborates on the network's architecture and the loss function for improving the performance based on the categories. The network's architecture is divided into object detection, classification, and segmentation, and each is discussed, including their contributions and limitations. There is also a brief indication of the simplification and optimization of the network for simplifying the architecture. Additionally, comparative performance results with a visualization of the final output of five existing techniques is elaborated. Finally, this review is concluded by pointing to particular challenges in lane marking detection, such as generalization problems and computational complexity. There is also a brief future direction for solving the issues, for instance, efficient neural network, Meta, and unsupervised learning.Entities:
Keywords: ADAS; DBSCAN; deep neural network (DNN); object detection; segmentation
Mesh:
Year: 2022 PMID: 36236780 PMCID: PMC9571289 DOI: 10.3390/s22197682
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Summary of lane detection techniques using DNN.
| Author | Deep Learning Technique | Categories | Achievement | Limitation |
|---|---|---|---|---|
| Single stage | ||||
| Li et al. [ | IANet | Segmentation | Suitable for two-class segmentation | High computation due to non-local features |
| Gurghian et al. [ | DeepLane | Classification | Fast detection with simple architecture | Application scenarios are limited |
| Van et al. [ | DLFNet | Segmentation | It does not have a predefined condition | Applicable for the fixed number of lanes |
| Ze et al. [ | RLaneNet | Regression | Capable of handling uncertain lane numbers without post-processing | The lane ordinate needs to be predefined. |
| Hou et al. [ | self-attention distillation | Segmentation | The strategy is more efficient | High computational complexity |
| Kim et al. [ | TLELane | Segmentation | Significant achievement on the small dataset | It can only detect the ego lane |
| Davy et al. [ | Lanenet | Segmentation | Capable of handling uncertain lane number | High computational complexity due to the H-Net |
| Xingang et al. [ | SCNN | Segmentation | Slice convolution for long lane | High computational complexity |
| Shriyash et al. [ | CooNet | Regression | Less computational network as does not require clustering | Applicable for the fixed number of lanes |
| Two-stage | ||||
| Ghafoorian et al. [ | EL-GAN | Segmentation | Can capture lane close to the label | Require a high number of parameters |
| Qin et al. [ | CNN-LSTM | Segmentation | Useful for the occlusion scene | Computational is complex |
| Zhang et al. [ | GLCNet | Segmentation | Capable of making efficient interlinks between subsections of the network | High computational complexity and difficulties in the training stage |
| Chen et al. [ | LMD based on VGG16 | Segmentation | Dilated convolution can expand the predicted field | The performance result is lower |
| Huang et al. [ | Spatial and temporal-based CNN | Object Detection | Spatial and temporal enrich the detection area | Complex architecture |
| Seokju et al. [ | VPGNet | Object Detection | Efficient in different environmental conditions | High computational complexity due to the post-processing |
| Huval et al. [ | EELane | Object Detection | Effective for the occlusion scene | It contains the perpetual prediction |
| Kim et al. [ | RANSAC | Classification | Overcome the limitations of traditional approaches | The structure of the network is not accurate enough |
Summary of the performances among various deep learning techniques.
| Authors | Detection Rate (%) | FPR (%) | FNR (%) | Recall (%) | Accuracy (%) | Precision (%) |
|---|---|---|---|---|---|---|
| Jongin et al. [ | 93 | - | - | - | - | - |
| Dan et al. [ | - | - | 10.03 | - | - | - |
| Soonhong et al. [ | 88.70 | - | - | - | - | - |
| Bei et al. [ | - | - | - | 92.8 | - | 95.49 |
| Xue et al. [ | - | 5.5 | - | - | - | - |
| Gurghian et al. [ | - | - | 99.9 | - | 98.96 | |
| He et al. [ | - | - | - | 93.80 | - | 95.49 |
| Kim et al. [ | 98 | - | - | - | - | - |
| Seokju et al. [ | 87 | - | - | 88 | - | - |
| Zhe et al. [ | - | 2.79 | 4.99 | 95.01 | - | 94.94 |
| Umar et al. [ | 99 | - | - | - | - | - |
| Davy et al. [ | - | 7.8 | 2.44 | - | 96.38 | - |
| Ghafoorian et al. [ | 4.12 | 3.36 | - | 96.39 | - | |
| Xingang et al. [ | - | 6.17 | 1.8 | - | 96.53 | - |
| Ze et al. [ | - | 3.9 | - | - | - | |
| Youjin et al. [ | 92.4 | - | - | - | - | |
| Xiaolong et al. [ | - | 1.41 | 4.53 | - | - | 95.65 |
| Wenjie et al. [ | - | 7.7 | - | - | - | |
| Tian et al. [ | - | - | - | 66.4 | 83.5 | |
| Huang et al. [ | - | - | - | 96.6 | - | 97.3 |
| Ye et al. [ | - | - | 5.17 | - | - | - |
| Chao et al. [ | - | - | - | 66 | 96.26 | 89 |
| Philion et al. [ | - | 7.2 | 4.5 | - | 95.2 | - |
| Azimi et al. [ | - | - | - | 85.95 | - | |
| Sun et al. [ | - | 2.0 | - | 96.4 | - | |
| Zhang et al. [ | 95.21 | - | - | - | - | - |
| Zou et al. [ | - | 4.24 | 1.84 | 95.8 | 97.2 | 85.7 |
| Nguyen et al. [ | - | - | - | - | 98.1 | - |
| Hou et al. [ | - | 6.02 | 2.05 | - | 96.64 | - |
| Fabio et al. [ | - | - | - | - | 76.53 | - |
| Lo et al. [ | - | - | - | - | - | - |
| Zang et al. [ | 82.44 | - | - | - | - | |
| Mamidala et al. [ | - | - | - | - | 96.1 | - |
| Liu et al. [ | - | - | - | - | 97.9 | - |
| Ko et al. [ | - | 2.94 | 2.63 | - | 96.7 | - |
Figure 1Different pre-processing technique (a) original, (b) cropped, (c) brighten (d) mirrored, (e) rotating and (f) perspective.
Figure 2The architecture of CNN based lane marking detection technique.
Figure 3Schematic diagram of VPGNet.
Figure 4Schematic diagram of spatial and temporal based LMD technique.
Figure 5Schematic diagram of DeepLane.
Figure 6Schematic diagram of Deep Convolution Neural Network based on the lane markings detector (LMD).
Figure 7Schematic diagram of EL-GAN.
Figure 8Schematic diagram of GLCNet.
Figure 9Schematic diagram of CNN-LSTM.
Figure 10Schematic diagram of SCCN.
Figure 11Schematic diagram of CooNet.
Figure 12Schematic diagram of Lanenet.
Figure 13Sample image frames of the Tusimple dataset.
Summary of lane detection techniques using DNN.
| Authors | DNN Method | FPS | FNS | Accuracy |
|---|---|---|---|---|
| Davy et al. [ | Lanenet | 7.8 | 2.44 | 96.38 |
| Ghafoorian et al. [ | EL-GAN | 4.12 | 3.36 | 96.39 |
| Xingang et al. [ | SCNN | 6.17 | 1.8 | 96.53 |
| Qin et al. [ | CNN-LSTM | 0.01416 | 0.0186 | 97.30 |
| Hou et al. [ | Self-attention distillation | 6.02 | 2.05 | 96.64 |
| Van et al. [ | ERFNet-DLSF | 0.1064 | 0.0983 | 93.38 |
Figure 14Predicted lane marking using DNN (a) Input (b) Lanenet (c) SCNN (d) CNN-LSTM (e) ERFNet-DLSF and (f) El-GAN.