Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
Implements models from scratch without abstraction layers and combines Flash Attention with FSDP for distributed training across 1-1000+ GPUs/TPUs. Supports parameter-efficient finetuning via LoRA/QLoRA with mixed-precision quantization (fp4/8/16/32) to reduce GPU memory requirements, while integrating with PyTorch Lightning and Lightning Cloud infrastructure for end-to-end pretraining, finetuning, and deployment workflows through declarative YAML recipes.
13,225 stars and 15,196 monthly downloads. Actively maintained with 5 commits in the last 30 days. Available on PyPI.
Stars
13,225
Forks
1,409
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 06, 2026
Monthly downloads
15,196
Commits (30d)
5
Dependencies
8
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Lightning-AI/litgpt"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
liangyuwang/Tiny-DeepSpeed
Tiny-DeepSpeed, a minimalistic re-implementation of the DeepSpeed library
microsoft/Text2Grad
🚀 Text2Grad: Converting natural language feedback into gradient signals for precise model...
catherinesyeh/attention-viz
Visualizing query-key interactions in language + vision transformers (VIS 2023)
FareedKhan-dev/Building-llama3-from-scratch
LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's...
huangjia2019/llm-gpt
From classic NLP to modern LLMs: building language models step by step. 异æ¥å›¾ä¹¦ï¼šã€Š GPT图解 å¤§æ¨¡åž‹æ˜¯æ€Žæ ·æž„å»ºçš„ã€‹-...