3 skills found
feifeibear / Long Context AttentionUSP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
InternLM / InternEvoInternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
Eugene29 / Megatron DeepSpeed ViTFork of Megatron-DeepSpeed with VIT bug fixes and model parallelisms (TP, TP-SP, Ulysses, etc) enabled for VIT. Pipeline Parallelism is not yet enabled.