Skip to content

[Robust Single-Stage Fully Sparse 3D Object Detection via Detachable Latent Diffusion, 2026, AAAI]

License

Notifications You must be signed in to change notification settings

QWTforGithub/RSDNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

54 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

RSDNet

This repo is the official project repository of the paper Robust Single-Stage Fully Sparse 3D Object Detection via Detachable Latent Diffusion.

  • [ arXiv ]
  • The code will be released shortly...
  • The code is being organied...[08/12/2025]

The Overall Framework

rsdnet

Citation

If you find our paper useful to your research (3D Detection or modeling multiple perturbations), please cite our work as an acknowledgment.

@article{qu2025robust,
  title={Robust Single-Stage Fully Sparse 3D Object Detection via Detachable Latent Diffusion},
  author={Qu, Wentao and Mei, Guofeng and Wang, Jing and Wu, Yujiao and Huang, Xiaoshui and Xiao, Liang},
  journal={arXiv preprint arXiv:2508.03252},
  year={2025}
}

Motivation of Modeling Multiple Perturbations in DDPMs

  • Enabling arbitrary distribution modeling would be a significant extension of DDPMs.
  • Since the prior distribution is defined as Gaussian, exiting DDPMs only can model Gaussion distributions.
  • However, for some tasks, Gaussian distributions maybe not the optimal choice, such as image restoration or super-resolution (often follow Laplacian distributions), fluorescence microscopy (often follow Passion distributions).
  • Modeling other distributions in DDPMs requires rederiving the posterior (the ground truth in DDPMs).
  • This is highly impractical due to the complex derivations involved.
  • The training of DDPMs essentially performs distribution matching.
  • Distribution matching aims to sample accurate intermediate variables, xt, xt-1,...,x0.
  • So, can we directly fit these intermediate variables and introduce perturbations to achieve diverse generation?
  • Thus, we introduce the notion of sample fitting to realize the noise-injection and denoising mechanisms in DDPMs.
  • This allows us to discard explicit distribution assumptions, since we directly fit the intermediate variables.
  • This removes the need to rederive the posterior, as intermediate variables are no longer sampled from a distribution but directly predicted.
  • Moreover, generation remains driven by stochastic perturbations introduced at inference, preserving diversity.
  • Our method is intended to broadly impact generative tasks, as it offers a general concept.
  • We plan to release extended versions in future work.
  • For the detailed idea, please refer to the uploaded PDF (the author’s group-meeting PPT).

Overview

Installation

Requirements

The following environment is recommended for running RSDNet (four or eight NVIDIA 4090 GPUs):

  • Ubuntu: 18.04 and above
  • gcc/g++: 7.5 and above
  • CUDA: 11.6 and above
  • PyTorch: 1.13.1 and above
  • python: 3.8 and above

Environment

  • Base environment
conda create -n dlf python=3.8 -y
conda activate dlf
pip install torch==1.13.1+cu116 torchvision==0.14.1+cu116 torchaudio==0.13.1 --extra-index-url https://download.pytorch.org/whl/cu116

cd RSDNet-main/envs
pip install -r requirements1.txt
pip install -r requirements2.txt
pip install -r requirements3.txt
pip install -r requirements4.txt
pip install -r requirements5.txt
pip install -r requirements6.txt
pip install -r requirements7.txt
pip install -r requirements8.txt
pip install -r requirements9.txt

cd RSDNet-main
python setup.py develop

Data Preparation

  • Please prefer to OpenPCDet for building datasets (nuScenes and Waymo Open).

nuScenes

  • Download the official nuScenes (or Baidu Disk(code:1111)) dataset (with Lidar Segmentation) and organize the downloaded files as follows:
    RSDNet-main/data
      ├── nuscenes
      │   │── v1.0-trainval (or v1.0-mini if you use mini)
      │   │   │── gt_database_10sweeps_withvelo
      │   │   │── lidarseg
      │   │   │── maps
      │   │   │── samples
      │   │   │── sweeps
      │   │   │── v1.0-mini
      │   │   │── v1.0-test
      │   │   │── v1.0-trainval
      │   │   │── nuscenes_10sweeps_withvelo_lidar.npy
      │   │   │── nuscenes_dbinfos_10sweeps_withvelo.pkl
      │   │   │── nuscenes_infos_10sweeps_train.pkl
      │   │   │── nuscenes_infos_10sweeps_val.pkl

Waymo Open

  • Download the official Waymo Open dataset (with Lidar Segmentation) and organize the downloaded files as follows:
    RSDNet-main/data
      ├── waymo
      │   │   │── ImageSets
      │   │   │── raw_data
      │   │   │── waymo_processed_data_v0_5_0
      │   │   │── waymo_processed_data_v0_5_0_gt_database_train_sampled_1
      │   │   │── waymo_processed_data_v0_5_0_gt_database_train_sampled_1_global.npy
      │   │   │── waymo_processed_data_v0_5_0_infos_train.pkl
      │   │   │── waymo_processed_data_v0_5_0_infos_val.pkl
      │   │   │── waymo_processed_data_v0_5_0_waymo_dbinfos_train_sampled_1.pkl

Model Zoo

Model Benchmark Only Training Data? Num GPUs Val NDS Val mAP log checkpoint
RSDNet nuScenes 4(bs=16) 71.9 69.3% train_log Link1, Link2
Model Benchmark Only Training Data? Num GPUs mAP/H_L1 mAP/H_L2 log checkpoint
RSDNet waymo 4(bs=8) 83.7/81.4 77.8/75.6 train_log Link1, Link2

Quick Start

  • Please prefer to OpenPCDet for training and testing.

Training

cd RSDNet-main/tools

# Training on nuScenes for multiple GPUs
bash scripts/train_nusc.sh 4
# Training on nuScenes for single GPU
python train_nusc.py

# Training on Waymo Open for multiple GPUs
bash scripts/train_waymo.sh 8
# Training on Waymo Open for single GPU
python train_waymo.py

Testing

cd RSDNet-main/tools

# Testing on nuScenes for multiple GPUs
bash scripts/test_nusc.sh 4
# Testing on nuScenes for single GPU
python test_nusc.sh

# Testing on waymo open for multiple GPUs
bash scripts/test_waymo.sh 8
# Testing on waymo open for single GPU
python test_waymo.sh 

About

[Robust Single-Stage Fully Sparse 3D Object Detection via Detachable Latent Diffusion, 2026, AAAI]

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published