SkillAgentSearch skills...

CSKD

Official code for Cumulative Spatial Knowledge Distillation for Vision Transformers (ICCV-2023) https://openaccess.thecvf.com/content/ICCV2023/html/Zhao_Cumulative_Spatial_Knowledge_Distillation_for_Vision_Transformers_ICCV_2023_paper.html

Install / Use

/learn @Zzzzz1/CSKD
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

This repo is the official implementation of the ICCV-2023 paper: Cumulative Spatial Knowledge Distillation for Vision Transformers.

Cumulative Spatial Knowledge Distillation

Framework

<div style="text-align:center"><img src=".github/cskd.png" width="80%" ></div>

Main Results

| | | |:---------------:|:-----------------:| | DeiT-Ti | 74.5 | | CSKD-Ti | 76.3 | | DeiT-S | 81.2 | | CSKD-S | 82.3 | | DeiT-B | 83.4 | | CSKD-B | 83.8 |

Installation

Environments:

  • Python 3.6
  • PyTorch 1.10.1
  • torchvision 0.11.2

Install the package:

sudo pip3 install -r requirements.txt

Getting started

Evaluate

python3 -m torch.distributed.launch --use_env --standalone --nnodes 1 --nproc_per_node 2 main.py --config configs/cskd_tiny.py --eval-only --ckpt {ckpt}

Train

python3 -m torch.distributed.launch --use_env --standalone --nnodes 1 --nproc_per_node 8 main.py --config configs/cskd_tiny.py

Citation

If this repo is helpful for your research, please consider citing the paper:

@inproceedings{zhao2023cumulative,
  title={Cumulative Spatial Knowledge Distillation for Vision Transformers},
  author={Zhao, Borui and Song, Renjie and Liang, Jiajun},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  pages={6146--6155},
  year={2023}
}

License

This repo is released under the MIT license. See LICENSE for details.

Acknowledgement

  • Thanks for DeiT. We build this repo based on DeiT.

Related Skills

View on GitHub
GitHub Stars15
CategoryContent
Updated3mo ago
Forks1

Languages

Python

Security Score

87/100

Audited on Dec 24, 2025

No findings