SkillAgentSearch skills...

SLGTformer

An Attention Based Approach to Sign Language Recognition | SOTA 2022 on WLASL Joints | https://arxiv.org/abs/2212.10746

Install / Use

/learn @neilsong/SLGTformer
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

SLGTformer: An Attention Based Approach to Sign Language Recognition

If you find this code useful for your research, consider citing:

@misc{https://doi.org/10.48550/arxiv.2212.10746,
  doi = {10.48550/ARXIV.2212.10746},
  url = {https://arxiv.org/abs/2212.10746},
  author = {Song, Neil},
  keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {SLGTformer: An Attention-Based Approach to Sign Language Recognition},
  publisher = {arXiv},
  year = {2022},
  copyright = {Creative Commons Attribution 4.0 International}
}

Environment Setup

Simply run conda env create -f environment.yml.

Data preparation

Please download and use the preprocessed skeleton data for WLASL by Skeleton Aware Multi-modal Sign Language Recognition. Please be sure to follow their rules and agreements when using the preprocessed data.

./download.sh

Pretrained models

Pretrained models are provided here.

Usage

Train WLASL:

./train.sh

Test:

./test.sh

Acknowledgements

SAM-SLR-v2

This code is based on SAM-SLR-v2. Huge thank you to the authors for open sourcing their code.

IRVL

Thank you to @yuxng for his advice and guidance throughout this project. Shout-out to his lab @IRVL for the RTX A5000s and all the fun conversations while models were training.

Related Skills

View on GitHub
GitHub Stars23
CategoryDevelopment
Updated26d ago
Forks2

Languages

Python

Security Score

75/100

Audited on Mar 9, 2026

No findings