SeaFormer
[ICLR 2023 & IJCV 2025] SeaFormer: Squeeze-enhanced Axial Transformer
Install / Use
/learn @fudan-zvg/SeaFormerREADME
Squeeze-enhanced axial Transformer
Paper
SeaFormer: Squeeze-enhanced Axial Transformer for Mobile Semantic Segmentation,
Qiang Wan, Zilong Huang, Jiachen Lu, Gang Yu, Li Zhang
ICLR 2023
SeaFormer++: Squeeze-enhanced Axial Transformer for Mobile Visual Recognition,
Qiang Wan, Zilong Huang, Jiachen Lu, Gang Yu, Li Zhang
IJCV 2025
This repository contains the official implementation of SeaFormer.
SeaFormer achieves superior trade-off between performance and latency
<div align="center"> <img width="1200", src="./latency.png"> </div>The overall architecture of Seaformer
<div align="center"> <img width="1200", src="./seaformer.png"> </div>The schematic illustration of the SeaFormer layer
<div align="center"> <img width="1200", src="./sea_attention.png"> </div>Model Zoo
Image Classification
Classification configs & weights see >>>here<<<.
- SeaFormer on ImageNet-1K
| Model | Size | Acc@1 | #Params (M) | FLOPs (G) | |------------------|:----:|:-----:|:-----------:|:---------:| | SeaFormer-Tiny | 224 | 68.1 | 1.8 | 0.1 | | SeaFormer-Small | 224 | 73.4 | 4.1 | 0.2 | | SeaFormer-Base | 224 | 76.4 | 8.7 | 0.3 | | SeaFormer-Large | 224 | 79.9 | 14.0 | 1.2 |
Semantic Segmentation
Segmentation configs & weights see >>>here<<<.
- SeaFormer on ADE20K
| Method | Backbone | Pretrain | Iters | mIoU(ss) | |--------------|------------------|-------------|-------|----------| | Light Head | SeaFormer-Tiny | ImageNet-1K | 160K | 36.5 | | Light Head | SeaFormer-Small | ImageNet-1K | 160K | 39.4 | | Light Head | SeaFormer-Base | ImageNet-1K | 160K | 41.9 | | Light Head | SeaFormer-Large | ImageNet-1K | 160K | 43.8 |
- SeaFormer on Cityscapes
| Method | Backbone | FLOPs | mIoU | |----------------|------------------|---------|----------| | Light Head(h) | SeaFormer-Small | 2.0G | 71.1 | | Light Head(f) | SeaFormer-Small | 8.0G | 76.4 | | Light Head(h) | SeaFormer-Base | 3.4G | 72.2 | | Light Head(f) | SeaFormer-Base | 13.7G | 77.7 |
BibTeX
@inproceedings{wan2023seaformer,
title={Seaformer: Squeeze-enhanced axial transformer for mobile semantic segmentation},
author={Wan, Qiang and Huang, Zilong and Lu, Jiachen and Gang, YU and Zhang, Li},
booktitle={International Conference on Learning Representations (ICLR)},
year={2023}
}
@article{wan2025seaformer++,
title={SeaFormer++: Squeeze-enhanced axial transformer for mobile visual recognition},
author={Wan, Qiang and Huang, Zilong and Lu, Jiachen and Yu, Gang and Zhang, Li},
journal={International Journal of Computer Vision (IJCV)},
year={2025}
}
Acknowledgment
Thanks to previous open-sourced repo:
TopFormer
mmsegmentation
pytorch-image-models
