FastBEV
base: https://github.com/Sense-GVT/Fast-BEV , delete time sequence,update mm releated ,add onnx export for tensorrt
Install / Use
/learn @cyn-liu/FastBEVREADME
FastBEV
Abstract
base: https://github.com/Sense-GVT/Fast-BEV
delete time sequence . you can add time seq in forward_3d refer to author's code.
update mmcv mmdet mmdet3d .... releted ,
add onnx export for tensorrt
fastbev-tiny ~= author's fastbev-m0. add neck fuse in m0
nuScenes is comming soon, wait few days (1-3day)
https://github.com/thfylsty/FastBEV-TensorRT
read install first for environment
TODO
[ ] author's data augment
[ ] evaluation fuction
DemoOnNuScenes
dataset convert
tools/create_data.sh
train
tools/dist_train.sh
in train.sh , we use fastbev-tiny.py ~= author's fastbev-m0
export
tools/dist_export.sh
test with nuscenes.pth
baiduPan:2cwz
Note: This pth model has not been trained well. There are also some abnormal predictions.
JUST FOR TEST EXPORT ONLY.
deploy
https://github.com/thfylsty/FastBEV-TensorRT
CustomDataset
how to convert to mm.pkl
refer to tools/dataset_converters/roadside_converter.py
other
update later maybe
用法
测试环境一
本地
- cuda 10.2
- cudnn 8.4.0
服务器
- cuda 11.7
- cudnn 8.4.0
基础
- Python = 3.8
- PyTorch = 1.10.0
- mmengine = 0.7.0
- mmcv = 2.0.0rc4 (>=2.0.0)
- mmdetection = 3.0.0rc6 (>=3.0.0)
- mmdetection3d = 1.1.0rc3 (>= 1.1.0)
测试环境二
服务器
- cuda 11.7
- cudnn 8.4.0
基础
- Python = 3.10
- PyTorch = 2.1.0
- mmengine = 0.7.0
- mmcv = 2.0.0rc4 (>=2.0.0)
- mmdetection = 3.0.0rc6 (>=3.0.0)
- mmdetection3d = 1.1.0rc3 (>= 1.1.0)
Getting Started
Evaluation
We also provide instructions for evaluating our pretrained models. Please download the checkpoints using the following script:
./tools/download_pretrained.sh
Then, you will be able to run:
torchpack dist-run -np 8 python tools/test.py [config file path] pretrained/[checkpoint name].pth --eval [evaluation type]
For example, if you want to evaluate the detection variant of BEVFusion, you can try:
torchpack dist-run -np 8 python tools/test.py configs/nuscenes/det/transfusion/secfpn/camera+lidar/swint_v0p075/convfuser.yaml pretrained/bevfusion-det.pth --eval bbox
While for the segmentation variant of BEVFusion, this command will be helpful:
torchpack dist-run -np 8 python tools/test.py configs/nuscenes/seg/fusion-bev256d2-lss.yaml pretrained/bevfusion-seg.pth --eval map
Training
We provide instructions to reproduce our results on nuScenes.
For example, if you want to train the camera-only variant for object detection, please run:
torchpack dist-run -np 8 python tools/train.py configs/nuscenes/det/centerhead/lssfpn/camera/256x704/swint/default.yaml --model.encoders.camera.backbone.init_cfg.checkpoint pretrained/swint-nuimages-pretrained.pth
For camera-only BEV segmentation model, please run:
torchpack dist-run -np 8 python tools/train.py configs/nuscenes/seg/camera-bev256d2.yaml --model.encoders.camera.backbone.init_cfg.checkpoint pretrained/swint-nuimages-pretrained.pth
For LiDAR-only detector, please run:
torchpack dist-run -np 8 python tools/train.py configs/nuscenes/det/transfusion/secfpn/lidar/voxelnet_0p075.yaml
For LiDAR-only BEV segmentation model, please run:
torchpack dist-run -np 8 python tools/train.py configs/nuscenes/seg/lidar-centerpoint-bev128.yaml
Acknowledgements
BEVFusion is based on mmdetection3d. It is also greatly inspired by the following outstanding contributions to the open-source community: LSS, BEVDet, TransFusion, CenterPoint, MVP, FUTR3D, CVT and DETR3D.
Please also check out related papers in the camera-only 3D perception community such as BEVDet4D, BEVerse, BEVFormer, M2BEV, PETR and PETRv2, which might be interesting future extensions to BEVFusion.
Related Skills
node-connect
347.0kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
107.8kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
347.0kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
347.0kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
