mosaicml/llm-foundry
LLM training code for Databricks foundation models
Implements end-to-end training, finetuning, evaluation, and inference pipelines with built-in support for efficiency techniques like Flash Attention and Mixture-of-Experts architectures. Integrates with Composer for distributed training optimization and MosaicML's platform for scalable workload orchestration, while supporting both HuggingFace and proprietary models (MPT, DBRX) from 125M to 132B parameters. Includes data preparation utilities for StreamingDataset format, inference export to ONNX/HuggingFace, and in-context learning evaluation on academic benchmarks.
4,397 stars and 4,165 monthly downloads. Available on PyPI.
Stars
4,397
Forks
584
Language
Python
License
Apache-2.0
Category
Last pushed
Oct 27, 2025
Monthly downloads
4,165
Commits (30d)
0
Dependencies
22
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/mosaicml/llm-foundry"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소
rllm-team/rllm
Pytorch Library for Relational Table Learning with LLMs.