Skip to content

[AAAI 2025] HGSFusion: Radar-Camera Fusion with Hybrid Generation and Synchronization for 3D Object Detection

License

Notifications You must be signed in to change notification settings

garfield-cpp/HGSFusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

HGSFusion: Radar-Camera Fusion with Hybrid Generation and Synchronization for 3D Object Detection

PWC Static Badge License: MIT

Abstract

Millimeter-wave radar plays a vital role in 3D object detection for autonomous driving due to its all-weather and all-lighting-condition capabilities for perception. However, radar point clouds suffer from pronounced sparsity and unavoidable angle estimation errors. To address these limitations, incorporating a camera may partially help mitigate the shortcomings. Nevertheless, the direct fusion of radar and camera data can lead to negative or even opposite effects due to the lack of depth information in images and low-quality image features under adverse lighting conditions. Hence, in this paper, we present the radar-camera fusion network with Hybrid Generation and Synchronization (HGSFusion), designed to better fuse radar potentials and image features for 3D object detection. Specifically, we propose the Radar Hybrid Generation Module (RHGM), which fully considers the Direction-Of-Arrival (DOA) estimation errors in radar signal processing. This module generates denser radar points through different Probability Density Functions (PDFs) with the assistance of semantic information. Meanwhile, we introduce the Dual Sync Module (DSM), comprising spatial sync and modality sync, to enhance image features with radar positional information and facilitate the fusion of distinct characteristics in different modalities. Extensive experiments demonstrate the effectiveness of our approach, outperforming the state-of-the-art methods in the VoD and TJ4DRadSet datasets by $6.53%$ and $2.03%$ in RoI AP and BEV AP, respectively.

Qualitative results

Visualization results on View of Delft validation set. Each row represents a frame. In each row, the images are the images with ground truth, detection results under the BEV of the proposed HGSFusion. Purple boxes represents the ground truth, and red boxes represents the detection results. Raw radar points are shown in blue and generated points are shown in orange. VoD Dataset Visualization results on the test set of TJ4DRadSet dataset under various lightning conditions. “Dark”, “Normal”, and “Shiny” are presented in different rows, respectively. In each row, the images are the images with ground truth and detection results under the BEV of the proposed HGSFusion. Green boxes denote the ground truth and red boxes represent the detection results. Raw radar points are shown in blue and generated points are shown in orange. TJ4DRadSet Dataset

Method

Overall framework Overall framework of the proposed HGSFusion. In the radar branch, the RHGM utilizes raw radar points and images to generate hybrid radar points (generated points, foreground points, and raw radar points shown in green, orange, and blue points, respectively). Then the hybrid radar points are encoded and passed through the radar backbone to produce radar BEV features. In the image branch, images are processed through image backbone and view transformation, producing image BEV features. Subsequently in DSM, the image and radar features undergo dual sync to obtain fused BEV features for object detection.

Citation

If you find our work helpful to your research, please consider citing:

@article{Gu_2025_AAAI,
  title={HGSFusion: Radar-Camera Fusion with Hybrid Generation and Synchronization for 3D Object Detection},
  author={Zijian Gu, Jianwei Ma, Yan Huang, Honghao Wei, Zhanye Chen, Hui Zhang, Wei Hong},
  journal={Proceedings of the AAAI Conference on Artificial Intelligence},
  year={2025}
}

Model Zoo

We offer the model on VoD and TJ4DRadset.

Datset Config Weight
VoD hgsfusion_vod.yaml Google Baidu
TJ4D hgsfusion_tj4d.yaml Google Baidu
  • The image backbone is pretrained on the COCO dataset, you can download from here and place the weight to checkpoints/deeplabv3_resnet101_coco-586e9e4e.pt

Environment

The requirements are the same as those of OpenPCDet

Install PyTorch 1.13 + CUDA 11.6:

conda create -n hgsfusion python=3.9.18
conda activate hgsfusion
pip install torch==1.13.0+cu116 -f https://download.pytorch.org/whl/torch_stable.html
pip install torchvision==0.14.0+cu116 -f https://download.pytorch.org/whl/torch_stable.html

Install other dependices:

pip install openmim
pip install mmcv==2.1.0
pip install mmdet==3.3.0

Compile CUDA extensions:

git clone https://github.com/garfield-cpp/HGSFusion.git
python setup.py develop
cd pcdet\ops\pillar_ops
python setup.py develop

Prepare Datset

  1. Download VoD and TJ4DRadset. Link the dataset to the folder under data/
mkdir data
ln -s /path/to/vod/dataset/ ./data/vod_radar_5frames
ln -s /path/to/tj4d/dataset/ ./data/tj4d
  1. You can download the hybrid radar points from Google or Baidu and unzip them to the dataset folder.

  2. (Optional) Or you can choose to generate hybrid radar points by yourself following here.

  3. Generate the pkl files of VoD and TJ4DRadset by replacing dataset_name to vod_dataset and tj4d_dataset respectively.

python pcdet.datasets.kitti.dataset_name
  1. Folder structure:
data
├── dataset_name
│   ├── ImageSets
│   ├── kitti_infos_test.pkl
│   ├── kitti_infos_train.pkl
│   ├── kitti_infos_trainval.pkl
│   ├── kitti_infos_val.pkl
│   ├── testing
│   └── training
|   |   ├── calib                               
|   |   ├── pose
|   |   ├── velodyne
|   |   ├── image_2
|   |   ├── mask_maskformer_with_label_k_1_gauss_k_4_uniform      
|   |   └── label_2                                        

Training and Evaluating

Train HGSFusion with 4 GPUs:

export CUDA_VISIBLE_DEVICES=0,1,2,3 bash ./tools/scripts/dist_train.sh 4

Single-GPU Evaluation:

python ./tools/test.py --cfg_file ./tools/cfgs/hgsfusion/hgsfusion_vod.yaml --ckpy ./path/to/your/ckpt

Acknowledgements

Many thanks to the open-source repositories:

About

[AAAI 2025] HGSFusion: Radar-Camera Fusion with Hybrid Generation and Synchronization for 3D Object Detection

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published