MapTR
[ICLR'23 Spotlight & ECCV'24 & IJCV'24] MapTR: Structured Modeling and Learning for Online Vectorized HD Map Construction
Install / Use
/learn @hustvl/MapTRREADME
Bencheng Liao<sup>1,2,3</sup> *, Shaoyu Chen<sup>1,3</sup> *, Yunchi Zhang<sup>1,3</sup> , Bo Jiang<sup>1,3</sup> ,Tianheng Cheng<sup>1,3</sup>, Qian Zhang<sup>3</sup>, Wenyu Liu<sup>1</sup>, Chang Huang<sup>3</sup>, Xinggang Wang<sup>1 :email:</sup>
<sup>1</sup> School of EIC, HUST, <sup>2</sup> Institute of Artificial Intelligence, HUST, <sup>3</sup> Horizon Robotics
(*) equal contribution, (<sup>:email:</sup>) corresponding author.
ArXiv Preprint (arXiv 2208.14437)
openreview ICLR'23, accepted as ICLR Spotlight
extended ArXiv Preprint MapTRv2 (arXiv 2308.05736), accepted to IJCV 2024
</div>News
Feb. 27th, 2025: Check out our latest work, DiffusionDrive, accepted to CVPR 2025! This study explores multi-modal end-to-end driving using diffusion models for real-time and real-world applications.Oct. 6th, 2024: MapTRv2 is accepted to IJCV 2024!Feb 20th, 2024: MapTRv2-based VADv2 is presented on arXiv paper project page.Aug. 31th, 2023: initial MapTRv2 is released at maptrv2 branch. Please rungit checkout maptrv2to use it.Aug. 14th, 2023: As required by many researchers, the code of MapTR-based map annotation framework (VMA) will be released at https://github.com/hustvl/VMA recently.Aug. 10th, 2023: We release MapTRv2 on Arxiv. MapTRv2 demonstrates much stronger performance and much faster convergence. To better meet the requirement of the downstream planner (like PDM), we introduce an extra semantic——centerline (using path-wise modeling proposed by LaneGAP). Code & model will be released in late August. Please stay tuned!May. 12th, 2023: MapTR now support various bevencoder, such as BEVFormer encoder and BEVFusion bevpool. Check it out!Apr. 20th, 2023: Extending MapTR to a general map annotation framework (paper, code), with high flexibility in terms of spatial scale and element type.Mar. 22nd, 2023: By leveraging MapTR, VAD (paper, code) models the driving scene as fully vectorized representation, achieving SoTA end-to-end planning performance!Jan. 21st, 2023: MapTR is accepted to ICLR 2023 as Spotlight Presentation!Nov. 11st, 2022: We release an initial version of MapTR.Aug. 31st, 2022: We released our paper on Arxiv. Code/Models are coming soon. Please stay tuned! ☕️
Introduction
<div align="center"><h4>MapTR/MapTRv2 is a simple, fast and strong online vectorized HD map construction framework.</h4></div>
High-definition (HD) map provides abundant and precise static environmental information of the driving scene, serving as a fundamental and indispensable component for planning in autonomous driving system. In this paper, we present Map TRansformer, an end-to-end framework for online vectorized HD map construction. We propose a unified permutation-equivalent modeling approach, i.e., modeling map element as a point set with a group of equivalent permutations, which accurately describes the shape of map element and stabilizes the learning process. We design a hierarchical query embedding scheme to flexibly encode structured map information and perform hierarchical bipartite matching for map element learning. To speed up convergence, we further introduce auxiliary one-to-many matching and dense supervision. The proposed method well copes with various map elements with arbitrary shapes. It runs at real-time inference speed and achieves state-of-the-art performance on both nuScenes and Argoverse2 datasets. Abundant qualitative results show stable and robust map construction quality in complex and various driving scenes.
Models
Results from the MapTRv2 paper

| Method | Backbone | Lr Schd | mAP| FPS| | :---: | :---: | :---: | :---: | :---: | MapTR | R18 | 110ep | 45.9 | 35.0| | MapTR | R50 | 24ep | 50.3 | 15.1| | MapTR | R50 | 110ep | 58.7|15.1| | MapTRv2 | R18 | 110ep | 52.3 | 33.7| | MapTRv2 | R50 | 24ep | 61.5 | 14.1| | MapTRv2 | R50 | 110ep | 68.7 | 14.1| | MapTRv2 | V2-99 | 110ep | 73.4 | 9.9|
Notes:
- FPS is measured on NVIDIA RTX3090 GPU with batch size of 1 (containing 6 view images).
- All the experiments are performed on 8 NVIDIA GeForce RTX 3090 GPUs.
Results from this repo.
MapTR
<div align="center"><h4> nuScenes dataset</h4></div>| Method | Backbone | BEVEncoder |Lr Schd | mAP| FPS|memory | Config | Download | | :---: | :---: | :---: | :---: | :---: | :---:|:---:| :---: | :---: | | MapTR-nano | R18 |GKT | 110ep |46.3 |35.0| 11907M (bs 24) |config |model / log | | MapTR-tiny | R50 | GKT |24ep | 50.0 |15.1| 10287M (bs 4) | config|model / log | | MapTR-tiny | R50 |GKT | 110ep | 59.3 |15.1| 10287M (bs 4)|config |model / log | | MapTR-tiny | Camera & LiDAR | GKT |24ep | 62.7 | 6.0 | 11858M (bs 4)|config |model / log | | MapTR-tiny | R50 | bevpool |24ep | 50.1 | 14.7 | 9817M (bs 4)|config |model / log | | MapTR-tiny | R50 | bevformer |24ep | 48.7 | 15.0 | 10219M (bs 4)|config |model / log | | MapTR-tiny<sup>+</sup> | R50 | GKT |24ep | 51.3 | 15.1 | 15158M (bs 4)|config |model / log | | MapTR-tiny<sup>+</sup> | R50 | bevformer |24ep | 53.3 | 15.0 | 15087M (bs 4)|config |model / log |
Notes:
- <sup>+</sup> means that we introduce temporal setting.
MapTRv2
Please git checkout maptrv2 and follow the install instruction to use following checkpoint
| Method | Backbone | BEVEncoder |Lr Schd | mAP| FPS|memory | Config | Download | | :---: | :---: | :---: | :---: | :---: | :---:|:---:| :---: | :---: | | MapTRv2| R50 |bevpool | 24ep | 61.4 |14.1| 19426M (bs 24) |config |model / log | | MapTRv2*| R50 |bevpool | 24ep | 54.3 |WIP| 20363M (bs 24) |config |model / log |
<div align="center"><h4> Argoverse2 dataset</h4></div>| Method | Backbone | BEVEncoder |Lr Schd | mAP| FPS|memory | Config | Download | | :---: | :---: | :---: | :---: | :---: | :---:|:---:| :---: | :---: | | MapTRv2| R50 |bevpool | 6ep | 64.3 |14.1| 20580 (bs 24) |config |model / [log](https://drive.google.
