| Literature DB >> 35898039 |
Guangzhen Cui1,2, Weili Zhang1, Yanqiu Xiao1,2, Lei Yao1,2, Zhanpeng Fang1,2.
Abstract
Cooperative perception, as a critical technology of intelligent connected vehicles, aims to use wireless communication technology to interact and fuse environmental information obtained by edge nodes with local perception information, which can improve vehicle perception accuracy, reduce latency, and eliminate perception blind spots. It has become a current research hotspot. Based on the analysis of the related literature on the Internet of vehicles (IoV), this paper summarizes the multi-sensor information fusion method, information sharing strategy, and communication technology of autonomous driving cooperative perception technology in the IoV environment. Firstly, cooperative perception information fusion methods, such as image fusion, point cloud fusion, and image-point cloud fusion, are summarized and compared according to the approaches of sensor information fusion. Secondly, recent research on communication technology and the sharing strategies of cooperative perception technology is summarized and analyzed in detail. Simultaneously, combined with the practical application of V2X, the influence of network communication performance on cooperative perception is analyzed, considering factors such as latency, packet loss rate, and channel congestion, and the existing research methods are discussed. Finally, based on the summary and analysis of the above studies, future research issues on cooperative perception are proposed, and the development trend of cooperative perception technology is forecast to help researchers in this field quickly understand the research status, hotspots, and prospects of cooperative perception technology.Entities:
Keywords: C-V2X; DSRC; IoV; autonomous driving; congestion control; cooperative perception; multi-sensor information fusion
Mesh:
Year: 2022 PMID: 35898039 PMCID: PMC9332497 DOI: 10.3390/s22155535
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1An overall diagram of the critical review on cooperative perception in the Internet of vehicles environment.
Figure 2The increasing amount of research on cooperative perception multi-sensor information fusion in the Internet of vehicles environment.
A comparison of the commonly employed sensors in self-driving cars, only sensors fusion, and combined V2X. The “√” symbol indicates that the sensor operates competently under the specific factor. The “~” indicates that the sensor performs reasonably well under the specific factor. The “×” indicates that the sensor does not operate well under specific factors relative to the other sensors.
| Factor | Camera | LiDAR | Radar | Only Fusion | Fusion and V2X |
|---|---|---|---|---|---|
| Range | ~ | ~ | √ | ~ | √ |
| Resolution | √ | ~ | × | √ | √ |
| Distance Accuracy | ~ | √ | √ | ~ | √ |
| Velocity | ~ | × | √ | √ | √ |
| Color Perception (e.g., | √ | × | × | √ | √ |
| Object Detection | ~ | √ | √ | ~ | √ |
| Object Classification | √ | ~ | × | √ | √ |
| Lane Detection | √ | × | × | √ | √ |
| Obstacle Edge Detection | √ | √ | × | √ | √ |
| Illumination Conditions | × | √ | √ | ~ | √ |
| Weather Conditions | × | ~ | √ | ~ | √ |
Figure 3Object detection from different angles based on image fusion.
Summary and analysis of image fusion research methods.
| Authors, | Key Research Points | Findings | Remarks |
|---|---|---|---|
| Lian et al. | Multiple roadside camera perception data mapping to form a global semantic description. | The detection time was increased by about 45%, and the detection accuracy was increased by about 10%. | Distributed interactive fusion deployment of sensors for a wider range of cooperative perception without increasing the time cost of computing. |
| Löhdefifink et al. | Used a lossy learning method for image compression to relieve the pressure on wireless communication channels. | Image compression requirements were high, and late fusion results of segmentation mask cascades were optimal. | The transmission of processed data can effectively reduce the load on the wireless channel. |
| Lv et al. | Based on the separation principle of static background and dynamic foreground, the dynamic foreground was extracted, and the static background and dynamic foreground were re-fused by a generative adversarial network. | The processing time of perceptual information was reduced to 27.7% of the original. | |
| Xiao et al. | A bird’s-eye view generated by integrating the perception information of other vehicles expanded the perception range, shared the processed image information, and reduced the network burden. | Solved the problem of obstacle occlusion and reduced the transmission of data volume. | Perception range will be affected by communication distance. |
| Sridhar1 | Utilized image feature point matching for data fusion to form vehicle cooperative perception with a common field of view. | Fusion of perception information from other vehicles and conversion to its own coordinate system. | Cooperative perception can effectively expand the perception range of vehicles. |
| Liu et al. | Used feature point matching to estimate geometric transformation parameters to solve perception blind spots in congestion. | The intersection over union value was increased by 2~3 times. | Effectively solved the obstacle occlusion, but ignored the problem of viewing angle. |
Figure 4Early and late fusion schemes for cooperative 3D object detection.
Summary of recent studies on point cloud fusion methods.
| Authors, | Key Research Points | Findings | Remarks |
|---|---|---|---|
| Chen et al. | Shared the original point cloud data for the first time, and analyzed the impact of communication cost and the robustness of positioning errors on cooperative perception. | Sparse point cloud negatively affects perception. | Data-level fusion. |
| Ye et al. | Fusion of raw sensor data from multiple vehicles to overcome occlusion and sensor resolution degradation with distance. | Fusing sensor data from multiple viewpoints improved perception accuracy and range. | Data-level fusion. |
| Chen et al. | A feature-level fusion scheme was proposed, and the tradeoffs between processing time, bandwidth usage, and detection performance were analyzed. | The detection accuracy within 20 m was improved by about 10%. | Feature-level fusion. |
| Wei et al. | Integrated the point cloud data of multiple objects, continuously perceiving the position of surrounding vehicles in cases of limited LiDAR perception and V2V communication failure. | Cooperative perception object detection was more stable than LiDAR-only and V2V-only methods. | Feature-level fusion. |
| Arnold et al. | Proposed early fusion and late fusion schemes of single-modal point cloud data to more accurately estimate the bounding box of the detection target. | The recall rate of cooperative perception target detection was as high as 95%. | The detection performance of data-level fusion was better than that of decision-level fusion, but the communication quality was poor. |
Figure 5A comparison between image data and point cloud data.
Summary and classification of research on image–point cloud fusion methods.
| Authors, | Key Research Points | Findings | Remarks |
|---|---|---|---|
| Jiang et al. | Used millimeter-wave radar to filter the target and map it to the image to obtain the region of interest, weighted the detection value and estimated value of the two, and improved the perception accuracy. | Effectively detected small targets in foggy weather. | Strong anti-interference ability. |
| Fu et al. | A fusion perception method of roadside camera millimeter-wave radar was proposed, and the Kalman filter was used to evaluate the pros and cons of the perception results. | Both horizontal and vertical had better detection results. | No actual deployment. |
| Wang et al. | Combined with real road scenes, filtered background objects detected by radar to achieve the automatic calibration of multiple sensors. | Fast and automatic acquisition of roadside perception fusion information. | Attempt to combine depth information to display detection results in 3D boxes. |
| Saito et al. | Projected the point cloud data to the pixel coordinate system of the next frame of the point cloud, performed 3D reconstruction, and improved the accuracy of target detection. | Improved target shape recovery rate and discernible distance. | Further adjustments to real-time models and panoramic cameras to expand the fusion range. |
| Duan et al. | An image–point cloud cooperative perception system was proposed, which sends the detected objects within the perception range to the vehicle. | Effectively extended the detection range. | A large amount of calculation and poor real-time performance. |
| Gu et al. | Utilized point cloud and image concatenation to form a point cloud single-modality mode and a point cloud–image multimodal mode fusion network. | Multimodality for more environmental changes. | Improved the detection accuracy of the road and had good real-time performance. |
Comparison of the advantages and disadvantages of three different fusion methods.
| Methodology | Advantages | Disadvantages | Conclusion |
|---|---|---|---|
| Image fusion | High resolution, richer perception information, good target classification effect, mature algorithm, low deployment cost. | The depth information of the target is insufficient, and it is greatly affected by light and adverse weather. | The image–point cloud fusion scheme has the best effect. |
| Point cloud fusion | High spatial resolution, rich 3D information, wide detection range, good positioning, and ranging effect. | Poor target classification effect, a large amount of data, easily affected by adverse weather, expensive. | |
| Image–point cloud fusion | Realizes the complementary advantages of images and point clouds, high resolution, high perception accuracy, richer information, and strong anti-interference. | A large amount of data and a complex algorithm. |
Figure 6C-V2X standardization and evaluation.
Figure 7V2X communication in a DSRC–cellular hybrid urban scenario.
Figure 8Application scenarios for 5G and C-V2X.
Figure 9Illustration of the cooperative perception conception.
Summary of the existing research on cooperative perception information-sharing communication technology according to market penetration rate and vehicle mobility.
| Factors | Authors, | Key Research Points | Findings |
|---|---|---|---|
| Vehicle Mobility | Zhu et al. | The network architecture of MEC and C-V2X fusion was proposed, which reduces the network transmission delay and improves the reliability of the perception system. | Distributed computing deployment can effectively reduce interaction delay. |
| Fukatsu et al. | Explored the requirements of different driving speeds for network data transmission. | The larger the bandwidth, the better the cooperative perception effect. | |
| Fukatsu et al. | Analyzed the data rate required to achieve cooperative perception at different driving speeds. | Derived the transmission data rate for safe driving at different driving speeds. | |
| Traffic Density and Market Penetration | Radovan et al. | Different sensor and communication equipment deployment schemes will effectively improve the scope of cooperative perception. | Different sensor combinations can make up for the lack of low permeability. |
| Li et al. | Analyzed the impact of market penetration on location and velocity estimates and forecasts. | When the market penetration rate is 50%, the estimated accuracy of vehicle positioning and speed is 80%-90%. | |
| Wang et al. | Discussed the capacity requirements for vehicle communication for cooperative perception under different traffic densities and market penetration rates. | V2I traffic from the CPM exchange is highest at about 50% penetration. |
Summary of the perceptual information-sharing strategies by type of shared information.
| Sharing Strategy | Authors, | Purposes | Findings | Remarks |
|---|---|---|---|---|
| CPM generation rules | Thandavarayan | Optimized the CPM generation strategy formulated by ETSI to reduce redundant information. | Dynamic CPM generation strategy. | Optimizing the CPM generation strategy can effectively reduce redundant information. |
| CPM value and freshness | Baldomero | Designed a context-based confirmation mechanism through which the transmitting vehicle can selectively request the confirmation of specified or critical broadcast information to reduce communication load. | Realized the correct reception of information through message response. | Transmitted critical sensory data, reducing communication load. |
| Higuchi | Decided whether to send the CPM policy by predicting the importance of the CPM to the receiver, reducing the communication load. | Leveraged value prediction networks and assessed the validity of information. | Shared perceptual information based on information importance and freshness. | |
| Aoki | Leveraged deep reinforcement learning to select data to transfer. | The detection accuracy was increased by 12%, and the packet reception rate was increased by 27%. | ||
| Rahal | Proposed enhancing the freshness of perceptual information to enhance the timeliness and accuracy of cooperative perception information. | Optimized information update rate. |
Summary of the possible impact of network performance on cooperative perception.
| Authors, | Key Research Points | Remarks |
|---|---|---|
| Liu et al. | Analyzed the impact of the analysis of factors affecting DSRC performance. | Communication distance and shelter are the main factors that cause the degradation of DSRC communication performance, and selective deployment of roadside equipment can effectively improve DSRC communication performance. |
| Bae et al. | Analyzed the impact of communication distance on packet reception rates in LoS and NLoS test scenarios. | Communication distance has a great influence on the reception rate of data packets. The greater the communication distance, the more serious the loss of packet reception rate. |
| Lee et al. | Analyzed the impact of PLR and delay on V2X data fusion. | By predicting data changes and using historical data, the accuracy of data fusion can be improved, and the detection accuracy is nearly 50% higher than that of lossy networks. |
| Xiong et al. | Evaluated the impact of latency and packet loss on the security of Internet of vehicles applications. | The higher the PLR, the lower the security. The smaller the initial speed, the lower the limit latency. |
| Thandavarayan et al. | The study investigated the impact of congestion control on cooperative perception using the DCC framework. | The combination of congestion control functions at the access and facility layers can improve the perception achieved with cooperative perception, ensure the timely transmission of the information, and significantly improve the object perception rate. |
| Günther et al. | Selected the best DCC variant and format of messages to maximize vehicle awareness. | The amount of data generated by cooperative perception can easily lead to channel congestion, resulting in too much old sensing information and reducing the accuracy of sensing information. |
| Furukawa et al. | Improved the vehicle position relationship and road structure to dynamically adjust the sensor data transmission rate method to improve the transmission rate of useful information. | Selecting high-probability vehicles to broadcast data and prioritizing data from other vehicles’ blind spots reduces radio traffic and enhances the real-time situational awareness of other vehicles. |
| Sepulcre et al. | Selected high-probability vehicles to broadcast and prioritize data from other vehicles’ blind spots, reducing radio traffic and enhancing real-time situational awareness of other vehicles. | Controlling the way the vehicle drops packets can reduce the flow of packets transmitted to the wireless channel, but the dropped packets are not transmitted, resulting in the lower performance of the application. |
Figure 10PDR levels for radios and applications with different DCC configurations when all packets have the same priority (DP2). Traffic density: 180 veh/km. (a) Radio level. (b) Application level.
Figure 11PDR levels for radios and applications with different DCC configurations when all packets have the same priority (DP2). Traffic density:120 veh/km. (a) Radio level. (b) Application level.
Figure 12Broadcast-style sharing of cooperative perception data.