학술논문

TL-4DRCF: A Two-Level 4-D Radar–Camera Fusion Method for Object Detection in Adverse Weather
Document Type
Periodical
Source
IEEE Sensors Journal IEEE Sensors J. Sensors Journal, IEEE. 24(10):16408-16418 May, 2024
Subject
Signal Processing and Analysis
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Robotics and Control Systems
Radar
Feature extraction
Point cloud compression
Cameras
Sensors
Radar detection
Three-dimensional displays
Autonomous driving
deep learning
millimeter-wave (MMW) radar
multimodal fusion
object detection
Language
ISSN
1530-437X
1558-1748
2379-9153
Abstract
In autonomous driving systems, cameras and light detection and ranging (LiDAR) are two common sensors for object detection. However, both sensors can be severely affected by adverse weather. With the development of radar technology, the emergence of the 4-D radar gives a more robust solution for sensor fusion strategies in 3-D object detection tasks. This study proposes a two-level 4-D radar and camera fusion model called TL-4DRCF, which performs a two-level fusion of the 4-D radar and camera information at the data and feature levels. In the data-level (DL) fusion stage, the radar point cloud is projected onto the image and fed as additional information to the image into the EarlyFusion-Net (EF-Net), which is the network designed for simultaneous extraction of point cloud and image features. In the feature-level fusion stage, the radar–camera alignment (RCA) module is proposed to accurately correlate point cloud voxels and pixel-level image features while consuming less inference time. The correlated features are used to predict the class and location of the object through a standard 3-D detection framework. The proposed TL-4DRCF was validated on the View-of-Delft (VoD) dataset and the VoD-Fog dataset performed by artificial fog processing. The experimental results show that the proposed model outperforms the baseline method PointPillars on the VoD dataset by 3.8% mAP and the LiDAR–camera-based method MVX-Net in the driving corridor area of the VoD-Fog dataset by 0.39% mAP.