Tencent/PatrickStar

PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.

48
/ 100
Emerging

Implements dynamic chunk-based memory scheduling to enable heterogeneous CPU-GPU training, allowing larger models (18B+ parameters) on fewer GPUs by intelligently offloading non-active model components to CPU memory. Integrates with PyTorch via a drop-in `initialize_engine()` API and supports multi-node distributed training with collective communication optimizations. Demonstrates 6x larger model capacity than DeepSpeed on equivalent hardware while maintaining faster training throughput.

779 stars.

No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

779

Forks

59

Language

Python

License

BSD-3-Clause

Last pushed

Nov 18, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Tencent/PatrickStar"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.