mosaicml/llm-foundry

LLM training code for Databricks foundation models

71
/ 100
Verified

Implements end-to-end training, finetuning, evaluation, and inference pipelines with built-in support for efficiency techniques like Flash Attention and Mixture-of-Experts architectures. Integrates with Composer for distributed training optimization and MosaicML's platform for scalable workload orchestration, while supporting both HuggingFace and proprietary models (MPT, DBRX) from 125M to 132B parameters. Includes data preparation utilities for StreamingDataset format, inference export to ONNX/HuggingFace, and in-context learning evaluation on academic benchmarks.

4,397 stars and 4,165 monthly downloads. Available on PyPI.

Maintenance 6 / 25
Adoption 18 / 25
Maturity 25 / 25
Community 22 / 25

How are scores calculated?

Stars

4,397

Forks

584

Language

Python

License

Apache-2.0

Last pushed

Oct 27, 2025

Monthly downloads

4,165

Commits (30d)

0

Dependencies

22

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/mosaicml/llm-foundry"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.