SkillAgentSearch skills...

MyTransformers

This repository provides a comprehensive library for parallel training and LoRA algorithm implementations, supporting multiple parallel strategies and a rich collection of LoRA variants. It serves as a flexible and efficient model fine-tuning toolkit for researchers and developers. Please contact hehn@mail.ustc.edu.cn for detailed information.

Install / Use

/learn @hhnqqq/MyTransformers
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

环境配置:

本仓库对环境的要求比较低:

  • transformers
  • liger_kernrl
  • 大于1.5.4的deepspeed
  • 大于2.2的torch

使用方法:

  • python setup.py install 配置环境

  • scripts/有一些训练脚本的示例

  • 调整脚本中的参数,具体参数的意思可以查看common/parser.py中的注释

  • 需要注意的是参数中的地址配置。本项目配置的是既可以通过指明地址来配置,也可以通过指明名字来配置

    • 比如配置ckpt path可以在脚本中直接写明地址
    • 也可以事先在MyTransfomers/paths.json文件中写入地址, 如下代码中配置了llama的tokenizer地址,那么通过llama的名字就可以取出该地址
  • 设置好参数之后运行该脚本即可启动训练

支持的功能:

  1. 使用liger_kernel减小训练显存使用,加速训练
  2. 使用deepspeed分布式训练,支持流水线并行,序列并行
  3. 支持使用transformers模型和tokenizer
  4. 支持20+ LoRA算法,架构清晰,易学习
  5. 支持灵活的optimizer设置
  6. 支持使用多种不同的attention implementation,支持对sdpa后端进行设置
  7. 支持多种不同的数据集
  8. 支持多节点训练
  9. 灵活的注册器机制

详细的文档:

MyTransformer使用文档

多机多卡训练

sacc

本仓库成果

  • [NeurIPS 2025] GoRA: Gradient-driven Adaptive Low Rank Adaptation
  • [EMNLP 2025 Findings] Biology-Instructions: A Dataset and Benchmark for Multi-Omics Sequence Understanding Capability of Large Language Models
  • [ICLR 2026] Gradient Intrinsic Dimensionality Alignment:Narrowing The Gap Between Low-Rank Adaptation and Full Fine-Tuning
  • [ICLR 2026] E²LoRA: Efficient and Effective Low-Rank Adaptation with Entropy-Guided Adaptive Sharing
  • [Under Review] Rethinking Multi-Omics LLMs from the Perspective of Omics-Encoding
  • [Under Review] A Unified Study of LoRA Variants: Taxonomy, Review, Codebase, and Empirical Evaluation
View on GitHub
GitHub Stars57
CategoryCustomer
Updated2d ago
Forks7

Languages

Python

Security Score

80/100

Audited on Mar 25, 2026

No findings