bentoml/BentoML
The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!
8,516 stars and 168,316 monthly downloads. Used by 4 other packages. Actively maintained with 15 commits in the last 30 days. Available on PyPI.
Stars
8,516
Forks
927
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 13, 2026
Monthly downloads
168,316
Commits (30d)
15
Dependencies
42
Reverse dependents
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mlops/bentoml/BentoML"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
nndeploy/nndeploy
一款简单易用和高性能的AI部署框架 | An Easy-to-Use and High-Performance AI Deployment Framework
kubeflow/trainer
Distributed AI Model Training and LLM Fine-Tuning on Kubernetes
cncf/llm-in-action
🤖 Discover how to apply your LLM app skills on Kubernetes!
ray-project/llms-in-prod-workshop-2023
Deploy and Scale LLM-based applications
SohamGovande/podplex
🦾💻🌐 distributed training & serverless inference at scale on RunPod