| Literature DB >> 35408157 |
Zhiqing Wei1, Fengkai Zhang1, Shuo Chang1, Yangyang Liu1, Huici Wu2, Zhiyong Feng1.
Abstract
With autonomous driving developing in a booming stage, accurate object detection in complex scenarios attract wide attention to ensure the safety of autonomous driving. Millimeter wave (mmWave) radar and vision fusion is a mainstream solution for accurate obstacle detection. This article presents a detailed survey on mmWave radar and vision fusion based obstacle detection methods. First, we introduce the tasks, evaluation criteria, and datasets of object detection for autonomous driving. The process of mmWave radar and vision fusion is then divided into three parts: sensor deployment, sensor calibration, and sensor fusion, which are reviewed comprehensively. Specifically, we classify the fusion methods into data level, decision level, and feature level fusion methods. In addition, we introduce three-dimensional(3D) object detection, the fusion of lidar and vision in autonomous driving and multimodal information fusion, which are promising for the future. Finally, we summarize this article.Entities:
Keywords: autonomous driving; data level fusion; decision level fusion; feature level fusion; lidar; object detection; radar and camera fusion; radar and vision fusion; review; survey
Mesh:
Year: 2022 PMID: 35408157 PMCID: PMC9003130 DOI: 10.3390/s22072542
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Process of mmWave radar and vision fusion.
Figure 2The organization of this article.
Autonomous driving datasets. “Y” indicates the existence of the sensing information in this dataset and “N” indicates the absence of the sensing information in this dataset.
| Dataset | Release Year | RGB Image | Radar | Lidar |
|---|---|---|---|---|
| Apolloscape | 2018 | Y | N | Y |
| KITTI | 2012 | Y | N | Y |
| Cityscapes | 2016 | Y | N | N |
| Waymo Open Dataset | 2019 | Y | N | Y |
| nuScense | 2019 | Y | Y | Y |
Autonomous driving sensor solutions of some manufacturers [30,31,32,33,34,35].
| Company | Autonomous Driving System | Sensor Configuration |
|---|---|---|
| Tesla | Autopilot | 8 cameras, 12 ultrasonic radars, mmWave radar |
| Baidu | Apollo | Lidar, mmWave radar, Camera |
| NIO | Aquila | Lidar, 11 cameras, 5 mmWave radars, 12 ultrasonic radars |
| Xpeng | XPILOT | 6 cameras, 2 mmWave radars, 12 ultrasonic radars |
| Audi | Traffic Jam Pilot | 6 cameras, 5 mmWave radars, 12 ultrasonic radars, Lidar |
| Mercedes Benz | Drive Pilot | 4 panoramic cameras, Lidar, mmWave radar |
Comparison of mmWave radar, lidar, and camera [36,37,38,39,40]. “1”–“6” denote the levels from “extremely low” to “extremely high”.
| Sensor Type | mmWave Radar | Lidar | Camera |
|---|---|---|---|
| Range resolution | 4 | 5 | 2 |
| Angle resolution | 4 | 5 | 6 |
| Speed detection | 5 | 4 | 3 |
| Detection accuracy | 2 | 5 | 6 |
| Anti-interference performance | 5 | 5 | 6 |
| Requirements for weather conditions | 1 | 4 | 4 |
| Operating hours | All weather | All weather | Depends on light conditions |
| Cost and processing overhead | 2 | 4 | 3 |
Figure 3Radar points are projected on the image and rendered into different colors.
Summary of the three fusion levels.
| Fusion Level | Advantages | Disadvantages |
|---|---|---|
| Data level | Minimum data loss and the highest reliability | Dependence on the number of radar points |
| Decision level | Making full use of sensor information | Modeling the joint probability density function of sensors is difficult |
| Feature level | Making full use of feature information and achieving best detection performance | Complicated computation and overhead of radar information transformation |
Summary of data level fusion.
| Reference | Contribution | ||
|---|---|---|---|
| ROI generation | [ | Using radar points to increase the speed of ROI generation. | |
| [ | Proposing the conclusion that distance determines the initial size of ROI. | ||
| [ | Extending ROI application to overtaking detection. | ||
| Object detection | Image preprocessing | [ | Using histogram equalization, grayscale variance and contrast normalization to preprocess the image. |
| [ | Image segmentation preprocessing with radar point as reference center. | ||
| Feature extraction | [ | Using features such as symmetry and shadow to extract vehicle contours. | |
| [ | Using Haar-like model for feature extraction. | ||
| Object classification | [ | Adaboost algorithm for object classification. | |
| [ | SVM for object classification. | ||
| [ | Neural network-based classifier for object classification. | ||
Figure 4Data level fusion.
Summary of decision level fusion.
| Reference | Contribution | ||
|---|---|---|---|
| Sensing information | Radar information | [ | The techniques involved in radar signal processing and what physical states can be obtained from radar information are analyzed. |
| Image | [ | Pedestrian detection using feature extraction combined with classifiers. | |
| [ | Detecting objects in depth images with MeanShift algorithm. | ||
| [ | An upgraded version of [ | ||
| [ | Using one-stage object detection algorithm YOLO for radar vision fusion object detection tasks. | ||
| Decision fusion | Based on Bayesian theory | [ | Proposing Bayesian programming to solve multi-sensor data fusion problems through probabilistic reasoning |
| [ | A dynamic fusion method based on Bayesian network is proposed to facilitate the addition of new sensors. | ||
| Based on Kalman filter | [ | Proposing a decision level fusion filter based on EKF framework. | |
| [ | The proposed fusion methon can track the object simultaneously in 3D space and 2D image plane. | ||
| [ | Functional equivalence of centralized and decentralized information fusion schemes is demonstrated. | ||
| Based on Dempster Shafer theory | [ | A decision level sensor fusion method based on Dempster-Shafer is proposed. | |
| Based on Radar validation | [ | Using radar detection results to validate visuals. | |
| [ | Using radar information to correct vehicle position information in real time to achieve object tracking. | ||
Figure 5Decision level fusion.
Summary of feature level fusion.
| Reference | Technology Features | |
|---|---|---|
| Fusion framework | [ | Based on SSD framework improvement, concatenation fusion is used. |
| [ | A fusion framework similar to YOLO structure is proposed named RVNet. | |
| [ | Proposing CRF-Net built on the VGG backbone network and RetinaNet, and the radar input branch is extended. | |
| [ | Join the radar branch based on the FCOS detection framework and embedded SAF module. | |
| Radar feature extraction | [ | Proposing a network named CMGGAN that can generate environmental images. |
| [ | Using a new radar feature description method called radar sparse image, the detected objects are presented as radar points. | |
| [ | Stretching the radar points in the radar sparse image vertically to supplement the height information. | |
| Feature fusion | [ | The fusion method of concatenation and element-wise addition is adopted. |
| [ | A feature fusion block named spatial attention fusion is proposed that uses attention mechanism. |
Figure 6Feature level fusion.