Tencent/PatrickStar
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
Implements dynamic chunk-based memory scheduling to enable heterogeneous CPU-GPU training, allowing larger models (18B+ parameters) on fewer GPUs by intelligently offloading non-active model components to CPU memory. Integrates with PyTorch via a drop-in `initialize_engine()` API and supports multi-node distributed training with collective communication optimizations. Demonstrates 6x larger model capacity than DeepSpeed on equivalent hardware while maintaining faster training throughput.
779 stars.
Stars
779
Forks
59
Language
Python
License
BSD-3-Clause
Category
Last pushed
Nov 18, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Tencent/PatrickStar"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
OpenMotionLab/MotionGPT
[NeurIPS 2023] MotionGPT: Human Motion as a Foreign Language, a unified motion-language...
OpenMotionLab/MotionGPT3
MotionGPT3: Human Motion as a Second Modality, a MoT-based framework for unified motion...
YerbaPage/DetectCodeGPT
Detection of LLM-Generated Codes [ICSE2025]
RobbenRibery/TuoTuo
TuoTuo is a Topic Modeling library for Researchers and Engineers
SmerkyG/gptcore
Fast modular code to create and train cutting edge LLMs