InternLM/xtuner

A Next-Generation Training Engine Built for Ultra-Large MoE Models

86
/ 100
Verified

Implements dropless FSDP training without expert parallelism for 200B+ MoE models, and supports 64k sequence lengths via memory optimization or DeepSpeed Ulysses sequence parallelism. Achieves higher throughput than traditional 3D parallelism for MoE scales above 200B, with optimized support for both NVIDIA GPUs and Ascend NPUs. Integrates with LMDeploy for inference and supports multimodal pre-training, supervised fine-tuning, and reinforcement learning algorithms like GRPO.

5,096 stars and 1,643 monthly downloads. Actively maintained with 72 commits in the last 30 days. Available on PyPI.

Maintenance 25 / 25
Adoption 17 / 25
Maturity 25 / 25
Community 19 / 25

How are scores calculated?

Stars

5,096

Forks

405

Language

Python

License

Apache-2.0

Last pushed

Mar 13, 2026

Monthly downloads

1,643

Commits (30d)

72

Dependencies

15

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/InternLM/xtuner"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.