OpenDriveVLA
[AAAI 2026] OpenDriveVLA: Towards End-to-end Autonomous Driving with Large Vision Language Action Model
Install / Use
/learn @DriveVLA/OpenDriveVLAREADME
OpenDriveVLA: Towards End-to-end Autonomous Driving with Large Vision Language Action Model
<h3 align="center"> <a href="https://drivevla.github.io/">Project Page</a> | <a href="https://arxiv.org/abs/2503.23463">arXiv</a> </h3>
Overview ✨
TODO List 📅
We will release the model code and checkpoints soon. Stay tuned! 🔥
- [x] Release environment setup
- [x] Release inference code
- [x] Release checkpoints
- [ ] Release training scripts
News 📢
2025/11/14Released the OpenDriveVLA 0.5B checkpoint on Hugging Face. 🌟2025/11/08OpenDriveVLA paper accepted by AAAI 2026. 🎉2025/08/10OpenDriveVLA model & inference code released. 🔥2025/04/01OpenDriveVLA paper is available on arXiv.2025/03/28We release the environment setup of OpenDriveVLA.- To make the dependencies of our OpenDriveVLA model [mmcv & mmdet3d] compatible with PyTorch 2.1.2 and support Transformers and Deepspeed, we selected specific versions and enhanced the source code accordingly. The resulting customized libraries are available in the
third_partyfolder.
- To make the dependencies of our OpenDriveVLA model [mmcv & mmdet3d] compatible with PyTorch 2.1.2 and support Transformers and Deepspeed, we selected specific versions and enhanced the source code accordingly. The resulting customized libraries are available in the
Getting Started 🌟
Citation 📝
If you find our project useful for your research, please consider citing our paper and codebase with the following BibTeX:
@misc{zhou2025opendrivevlaendtoendautonomousdriving,
title={OpenDriveVLA: Towards End-to-end Autonomous Driving with Large Vision Language Action Model},
author={Xingcheng Zhou and Xuyuan Han and Feng Yang and Yunpu Ma and Volker Tresp and Alois Knoll},
year={2025},
eprint={2503.23463},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2503.23463},
}
