SkillAgentSearch skills...

Unidet3d

[AAAI2025] UniDet3D: Multi-dataset Indoor 3D Object Detection

Install / Use

/learn @filaPro/Unidet3d
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

UniDet3D: Multi-dataset Indoor 3D Object Detection

News:

  • :fire: December, 2024. UniDet3D is now accepted at AAAI 2025.
  • :fire: September, 2024. UniDet3D is state-of-the-art in 6 indoor benchmarks: <br> ScanNet PWC <br> ARKitScenes PWC <br> S3DIS PWC <br> MultiScan PWC <br> 3RScan PWC <br> ScanNet++ PWC.

This repository contains an implementation of UniDet3D, a multi-dataset indoor 3D object detection method introduced in our paper:

UniDet3D: Multi-dataset Indoor 3D Object Detection<br> Maksim Kolodiazhnyi, Anna Vorontsova, Matvey Skripkin, Danila Rukhovich, Anton Konushin <br> Artificial Intelligence Research Institute <img src="https://github.com/user-attachments/assets/c6304076-153a-4e3b-b58c-db522f7f78fe" width="50"/> <br> https://arxiv.org/abs/2409.04234

Installation

For convenience, we provide a Dockerfile. This implementation is based on mmdetection3d framework v1.1.0. If not using Docker, please follow getting_started.md for the installation instructions.

Getting Started

Please see test_train.md for some basic usage examples.

Data Preprocessing

UniDet3D is trained and tested using 6 datasets: ScanNet, ARKitScenes, S3DIS, MultiScan, 3RScan, and ScanNet++. Preprocessed data can be found at our Hugging Face. Download each archive, unpack, and move into the corresponding directory in data. Please comply with the license agreement before downloading the data.

Alternatively, you can preprocess the data by youself. Training data for 3D object detection methods that do not requires superpoints, e.g. TR3D or FCAF3D, can be prepared according to the instructions.

Superpoints for ScanNet and MultiScan are provided as a part of the original annotation. For the rest datasets, you can either download pre-computed superpoints at our Hugging Face, or compute them using superpoint_transformer.

Training

Before training, please download the backbone checkpoint and save it under work_dirs/tmp.

To train UniDet3D on 6 datasets jointly, simply run the training script:

python tools/train.py configs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes.py

UniDet3D can also be trained on individual datasets, e.g., we provide a config for training using ScanNet solely.

Testing

To test a trained model, you can run the testing script:

python tools/test.py configs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes.py \
    work_dirs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes/epoch_1024.pth

UniDet3D can also be tested on individual datasets. To this end, simply remove the unwanted datasets from val_dataloader.dataset.datasets in the config file.

Visualization

To visualize ground truth and predicted boxes, run the testing script with additional arguments:

python tools/test.py configs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes.py \
    work_dirs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes/latest.pth --show \
    --show-dir work_dirs/unidet3d_1xb8_scannet_s3dis_multiscan_3rscan_scannetpp_arkitscenes

You can also set score_thr in configs to 0.3 for better visualizations.

Trained Model

Please refer to the UniDet3D checkpoint and log file. The corresponding metrics are given below (they might slightly deviate from the values reported in the paper due to the randomized training/testing procedure).

| Dataset | mAP<sub>25</sub> | mAP<sub>50</sub> | |:-----------:|:-----------------:|:-----------------:| | ScanNet | 77.0 | 65.9 | | ARKitScenes | 60.1 | 47.2 | | S3DIS | 76.7 | 65.3 | | MultiScan | 62.6 | 52.3 | | 3RScan | 63.6 | 44.9 | | ScanNet++ | 24.0 | 16.8 |

Predictions Example

<p align="center"> <img src="https://github.com/user-attachments/assets/bb535823-cc9b-4482-a1b6-d10cf74c9389" alt="UniDet3D predictions"/> </p>

Citation

If you find this work useful for your research, please cite our paper:

@inproceedings{kolodiazhnyi2025unidet3d,
  title={Unidet3d: Multi-dataset indoor 3d object detection},
  author={Kolodiazhnyi, Maksim and Vorontsova, Anna and Skripkin, Matvey and Rukhovich, Danila and Konushin, Anton},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={39},
  number={4},
  pages={4365--4373},
  year={2025}
}
View on GitHub
GitHub Stars168
CategoryDevelopment
Updated3d ago
Forks13

Languages

Python

Security Score

85/100

Audited on Apr 2, 2026

No findings