OpenCOOD
[ICRA 2022] An opensource framework for cooperative detection. Official implementation for OPV2V.
Install / Use
/learn @DerrickXuNu/OpenCOODREADME
OpenCOOD is an <strong>Open</strong> <strong>COO</strong>perative <strong>D</strong>etection framework for autonomous driving. It is also the official implementation of the <strong> ICRA 2022 </strong> paper OPV2V.
<p align="center"> <img src="images/demo1.gif" width="600" alt="" class="img-responsive"> <img src="images/camera_demo.gif" width="600" alt="" class="img-responsive"> </p>News:
- 12/28/2022: OpenCOOD now support multi-gpu training.
- 12/21/2022: V2XSet (ECCV2022) is supported by OpenCOOD now!
- 12/16/2022: Both spconv 1.2.1 and spconv 2.x are supported!
- 12/04/2022: The log replay tool for OPV2V is online now! With this toolbox, you can 100% replay all the events in the offline dataset and add/change any sensors/groundtruth you want to explore the tasks that the origin dataset do not support. Check here to see more details.
- 09/15/2022: So far OpenCOOD has supported several top conference papers, including ECCV,ICRA,CoRL,NeurIPS,WACV! The bottom of this project page lists the detailed information.
Features
-
Provide easy data API for multiple popular multi-agent perception dataset:
-
Provide APIs to allow users use different sensor modalities
- [x] LiDAR APIs
- [ ] Camera APIs
- [ ] Radar APIs
-
Provide multiple SOTA 3D detection backbone:
- [X] PointPillar
- [X] Pixor
- [X] VoxelNet
- [X] SECOND
-
Support multiple sparse convolution versions
- [X] Spconv 1.2.1
- [X] Spconv 2.x
-
Support SOTA multi-agent perception models:
- [x] Attentive Fusion [ICRA2022]
- [x] Cooper [ICDCS]
- [x] F-Cooper [SEC2019]
- [x] V2VNet [ECCV2022]
- [x] CoAlign (fusion only) [ICRA2023]
- [x] FPV-RCNN [RAL2022]
- [ ] DiscoNet [NeurIPS2021]
- [x] V2X-ViT [ECCV2022]
- [x] CoBEVT [CoRL2022]
- [ ] AdaFusion [WACV2023]
- [x] Where2comm [NeurIPS2022]
- [x] V2VAM [TIV2023]
-
Provide a convenient log replay toolbox for OPV2V dataset. Check here to see more details.
Data Downloading
All the data can be downloaded from UCLA BOX. If you have a good internet, you can directly
download the complete large zip file such as train.zip. In case you suffer from downloading large files, we also split each data set into small chunks, which can be found
in the directory ending with _chunks, such as train_chunks. After downloading, please run the following command to each set to merge those chunks together:
cat train.zip.part* > train.zip
unzip train.zip
Installation
Please refer to data introduction and installation guide to prepare data and install OpenCOOD. To see more details of OPV2V data, please check our website.
Quick Start
Data sequence visualization
To quickly visualize the LiDAR stream in the OPV2V dataset, first modify the validate_dir
in your opencood/hypes_yaml/visualization.yaml to the opv2v data path on your local machine, e.g. opv2v/validate,
and the run the following commond:
cd ~/OpenCOOD
python opencood/visualization/vis_data_sequence.py [--color_mode ${COLOR_RENDERING_MODE}]
Arguments Explanation:
color_mode: str type, indicating the lidar color rendering mode. You can choose from 'constant', 'intensity' or 'z-value'.
Train your model
OpenCOOD uses yaml file to configure all the parameters for training. To train your own model from scratch or a continued checkpoint, run the following commonds:
python opencood/tools/train.py --hypes_yaml ${CONFIG_FILE} [--model_dir ${CHECKPOINT_FOLDER} --half]
Arguments Explanation:
hypes_yaml: the path of the training configuration file, e.g.opencood/hypes_yaml/second_early_fusion.yaml, meaning you want to train an early fusion model which utilizes SECOND as the backbone. See Tutorial 1: Config System to learn more about the rules of the yaml files.model_dir(optional) : the path of the checkpoints. This is used to fine-tune the trained models. When themodel_diris given, the trainer will discard thehypes_yamland load theconfig.yamlin the checkpoint folder.half(optional): If set, the model will be trained with half precision. It cannot be set with multi-gpu training togetger.
To train on multiple gpus, run the following command:
CUDA_VISIBLE_DEVICES=0,1,2,3 python -m torch.distributed.launch --nproc_per_node=4 --use_env opencood/tools/train.py --hypes_yaml ${CONFIG_FILE} [--model_dir ${CHECKPOINT_FOLDER}]
Test the model
Before you run the following command, first make sure the validation_dir in config.yaml under your checkpoint folder
refers to the testing dataset path, e.g. opv2v_data_dumping/test.
python opencood/tools/inference.py --model_dir ${CHECKPOINT_FOLDER} --fusion_method ${FUSION_STRATEGY} [--show_vis] [--show_sequence]
Arguments Explanation:
model_dir: the path to your saved model.fusion_method: indicate the fusion strategy, currently support 'early', 'late', and 'intermediate'.show_vis: whether to visualize the detection overlay with point cloud.show_sequence: the detection results will visualized in a video stream. It can NOT be set withshow_visat the same time.global_sort_detections: whether to globally sort detections by confidence score. If set to True, it is the mainstream AP computing method, but would increase the tolerance for FP (False Positives). OPV2V paper does not perform the global sort. Please choose the consistent AP calculation method in your paper for fair comparison.
The evaluation results will be dumped in the model directory.
Benchmark and model zoo
Results on OPV2V LiDAR-track (AP@0.7 for no-compression/ compression)
| | Spconv Version| Backbone | Fusion Strategy | Bandwidth (Megabit), <br/> before/after compression| Default Towns |Culver City| Download | |--------------------| -------|-------- | --------------- | --------------- | ------------- |-----------| -------- | | Naive Late | 1.2.1 | PointPillar | Late | 0.024/0.024 | 0.781/0.781 | 0.668/0.668 | url | | Cooper | 1.2.1 | PointPillar | Early | 7.68/7.68 | 0.800/x | 0.696/x | url | | Attentive Fusion | 1.2.1 | PointPillar | Intermediate | 126.8/1.98 | 0.815/0.810 | 0.735/0.731 | url | | F-Cooper | 1.2.1 | PointPillar | Intermediate | 72.08/1.12 | 0.790/0.788 | 0.728/0.726 | url | | V2VNet | 1.2.1 | PointPillar | Intermediate | 72.08/1.12 | 0.822/0.814 | 0.734/0.729 | url | | CoAlign | 1.2.1 | PointPillar | Intermediate | 72.08/2.24 | 0.833/0.806 | 0.760/ 0.750 | url | FPV-RCNN | 1.2.1 | PV-RCNN | Intermediate(2 stage) | 0.24/0.24 | 0.820/0.820 | 0.763/0.763 | url | | V2VAM | 1.2.1 | PointPillar | Intermediate | x/x | 0.860/0.860 | 0.813/0.791 | url | | CoBEVT | 2.0 | PointPillar | Intermediate | 72.08/1.12 | 0.861/0.836 | 0.773/0.730 | url | | Naive Late | 1.2.1 | VoxelNet | Late | 0.024/0.024 | 0.738/0.738 | 0.588/0.588
