bentoml/OpenLLM
Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.
12,161 stars and 4,730 monthly downloads. Available on PyPI.
Stars
12,161
Forks
803
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 09, 2026
Monthly downloads
4,730
Commits (30d)
0
Dependencies
15
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/bentoml/OpenLLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)