SkillAgentSearch skills...

Rein

[CVPR 2024] Official implement of <Stronger, Fewer, & Superior: Harnessing Vision Foundation Models for Domain Generalized Semantic Segmentation>

Install / Use

/learn @w1oves/Rein
About this skill

Quality Score

0/100

Supported Platforms

Zed

README

[CVPR 2024] Stronger, Fewer, & Superior: Harnessing Vision Foundation Models for Domain Generalized Semantic Segmentation

zhixiang wei<sup>1</sup>, lin chen<sup>2</sup>, et al. <br /> <sup>1</sup> University of Science of Techonology of China <sup>2</sup> Shanghai AI Laboratory

PWC <br /> PWC <br /> PWC <br /> Project page: https://zxwei.site/rein

Paper: https://arxiv.org/pdf/2312.04265.pdf

Rein is a efficient and robust fine-tuning method, specifically developed to effectively utilize Vision Foundation Models (VFMs) for Domain Generalized Semantic Segmentation (DGSS). It achieves SOTA on Cityscapes to ACDC, and GTAV to Cityscapes+Mapillary+BDD100K. Using only synthetic data, Rein achieved an mIoU of 78.4% on Cityscapes validation set! Using only the data from the Cityscapes training set, we achieved an average mIoU of 77.6% on ACDC test set! Rein Framework

Visualization

Trained on Cityscapes, Rein generalizes to unseen driving scenes and cities: Nighttime Shanghai, Foggy Countryside, and Rainy Hollywood.

![night]

![rain]

![fog]

🔥 News!

  • 🔥 Welcome to check out our latest work: Rein++: Efficient Generalization and Adaptation for Semantic Segmentation with Vision Foundation Models!

  • 🔥 Delighted to announce that ours work HQCLIP: Leveraging Vision-Language Models to Create High-Quality Image-Text Datasets and CLIP Models were accepted by ICCV 2025!

  • 🔥 We warmly congratulate SoMA (https://ysj9909.github.io/SoRA.github.io/) for receiving a CVPR 2025 Highlight! Built upon the Rein codebase, SoMA demonstrates outstanding semantic segmentation performance and has even successfully accomplished object detection tasks!

  • 🔥 To facilitate users in integrating reins into their own projects, we provide a simplified version of reins: simple_reins. With this version, users can easily use reins as a feature extractor. (Note: This version has removed features related to mask2former)

  • We have uploaded the config for ResNet and ConvNeXt.

  • We have uploaded the checkpoint and config for +1/16 of Cityscapes training set, and it get 82.5% on the Cityscapes validation set!

  • Rein is accepted in CVPR2024!

  • Using only the data from the Cityscapes training set, we achieved an average mIoU of 77.56% on the ACDC test set! This result ranks first in the DGSS methods on the ACDC benchmark! Checkpoint is avaliable at release.

  • Using only synthetic data (UrbanSyn, GTAV, and Synthia), Rein achieved an mIoU of 78.4% on Cityscapes! Checkpoint is avaliable at release.

Performance Under Various Settings (DINOv2).

|Setting |mIoU |Config|Log & Checkpoint| |-|-|-|-| |GTAV $\rightarrow$ Cityscapes|66.7|config|log & checkpoint |+Synthia $\rightarrow$ Cityscapes|72.2|config|log & checkpoint| |+UrbanSyn $\rightarrow$ Cityscapes|78.4|config|log & checkpoint| |+1/16 of Cityscapes training $\rightarrow$ Cityscapes|82.5|config| log & checkpoint |GTAV $\rightarrow$ BDD100K|60.0|config|log & checkpoint |Cityscapes $\rightarrow$ ACDC|77.6|config|log & checkpoint |Cityscapes $\rightarrow$ Cityscapes-C|60.0|config|log & checkpoint

Performance For Various Backbones (Trained on GTAV).

|Setting |Pretraining|Citys. mIoU |Config|Log & Checkpoint| |-|-|-|-|-| |ResNet50 |ImageNet1k|49.1|config|log & checkpoint |ResNet101 |ImageNet1k|45.9|config| log & checkpoint |ConvNeXt-Large |ImageNet21k| 57.9|config|log & checkpoint |ViT-Small |DINOv2|55.3|config|log & checkpoint |ViT-Base |DINOv2|64.3|config|log & checkpoint |CLIP-Large | OPENAI | 58.1 | config|log & checkpoint |SAM-Huge |SAM| 59.2 | config|log & checkpoint |EVA02-Large |EVA02|67.8| config | log & checkpoint

Citation

If you find our code or data helpful, please cite our paper:

@InProceedings{Wei_2024_CVPR,
    author    = {Wei, Zhixiang and Chen, Lin and Jin, Yi and Ma, Xiaoxiao and Liu, Tianle and Ling, Pengyang and Wang, Ben and Chen, Huaian and Zheng, Jinjin},
    title     = {Stronger Fewer \& Superior: Harnessing Vision Foundation Models for Domain Generalized Semantic Segmentation},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2024},
    pages     = {28619-28630}
}

Try and Test

Experience the demo: Users can open [

View on GitHub
GitHub Stars401
CategoryDevelopment
Updated15d ago
Forks22

Languages

Python

Security Score

95/100

Audited on Mar 23, 2026

No findings