LLM Docker Deployments LLM Tools
Docker containerization and deployment solutions for running LLMs, inference servers, and related AI services locally or on networks. Does NOT include general containerization tools, Kubernetes orchestration, or non-LLM Docker projects.
There are 150 llm docker deployments tools tracked. 1 score above 70 (verified tier). The highest-rated is containers/ramalama at 82/100 with 2,640 stars. 4 of the top 10 are actively maintained.
Get all 150 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-docker-deployments&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving... |
|
Verified |
| 2 |
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services... |
|
Established |
| 3 |
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally |
|
Established |
| 4 |
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints.... |
|
Established |
| 5 |
vtuber-plan/olah
Self-hosted huggingface mirror service. 自建huggingface镜像服务。 |
|
Established |
| 6 |
foldl/chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your... |
|
Established |
| 7 |
quantalogic/qllm
QLLM: A powerful CLI for seamless interaction with multiple Large Language... |
|
Established |
| 8 |
eastriverlee/LLM.swift
LLM.swift is a simple and readable library that allows you to interact with... |
|
Established |
| 9 |
varunvasudeva1/llm-server-docs
End-to-end documentation to set up your own local & fully private LLM server... |
|
Established |
| 10 |
dingodb/dingospeed
dingospeed is a self-hosted huggingface mirror service |
|
Established |
| 11 |
Scottcjn/llama-cpp-power8
AltiVec/VSX optimized llama.cpp for IBM POWER8 |
|
Established |
| 12 |
lordmathis/llamactl
Unified management and routing for llama.cpp, MLX and vLLM models with web dashboard. |
|
Established |
| 13 |
sangyuxiaowu/LLamaWorker
LLamaWorker is a HTTP API server developed based on the LLamaSharp project.... |
|
Emerging |
| 14 |
liltom-eth/llama2-webui
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere... |
|
Emerging |
| 15 |
jlonge4/local_llama
This repo is to showcase how you can run a model locally and offline, free... |
|
Emerging |
| 16 |
France-Travail/happy_vllm
A REST API for vLLM, production ready |
|
Emerging |
| 17 |
FarisZahrani/llama-cpp-py-sync
Auto-synced CFFI ABI python bindings for llama.cpp with prebuilt wheels... |
|
Emerging |
| 18 |
ADT109119/llamacpp-distributed-inference
一個基於 llama.cpp 的分佈式 LLM 推理程式,讓您能夠利用區域網路內的多台電腦協同進行大型語言模型的分佈式推理,使用 Electron... |
|
Emerging |
| 19 |
timhagel/MeloTTS-Docker-API-Server
A docker image to access MeloTTS through API calls |
|
Emerging |
| 20 |
hitomi-team/sukima
A ready-to-deploy container for implementing an easy to use REST API to... |
|
Emerging |
| 21 |
Mcourtyard/m-courtyard
M-Courtyard: Local AI Model Fine-tuning Assistant for Apple Silicon.... |
|
Emerging |
| 22 |
feiyun0112/Local-LLM-Server
quick way to build a private large language model server and provide... |
|
Emerging |
| 23 |
cdrage/containerfiles
Containerfiles including AI, game servers, bootc and even a rickroll. |
|
Emerging |
| 24 |
icppWorld/icpp_llm
on-chain LLMs |
|
Emerging |
| 25 |
wsmlby/homl
The easiest & fastest way to run LLMs in your home lab |
|
Emerging |
| 26 |
gsuuon/ad-llama
Structured inference with Llama 2 in your browser |
|
Emerging |
| 27 |
ashleykleynhans/runpod-worker-oobabooga
RunPod Serverless Worker for Oobabooga Text Generation API for LLMs |
|
Emerging |
| 28 |
b-data/jupyterlab-mojo-docker-stack
(GPU accelerated) Multi-arch (linux/amd64, linux/arm64/v8) JupyterLab... |
|
Emerging |
| 29 |
john-rocky/EdgeLLM
Simple LLM package for ios devices. |
|
Emerging |
| 30 |
nicksavarese/allora-ios
An iOS Keyboard Extension that allows for interacting with LLMs directly... |
|
Emerging |
| 31 |
DanielZhangyc/RLLM
LLM powered RSS reader |
|
Emerging |
| 32 |
ai-action/ollama-action
🦙 Run Ollama large language models (LLMs) with GitHub Actions. |
|
Emerging |
| 33 |
abundant-ai/oddish
Run Harbor tasks in the cloud |
|
Emerging |
| 34 |
ruska-ai/llm-server
🤖 Open-source LLM server (OpenAI, Ollama, Groq, Anthropic) with support for... |
|
Emerging |
| 35 |
BlackTechX011/Ollama-in-GitHub-Codespaces
Learn all how to run Ollama in GitHub Codespaces for free |
|
Emerging |
| 36 |
Flowm/llm-stack
Docker compose config for local and hosted llms with multiple chat interfaces |
|
Emerging |
| 37 |
persys-ai/persys
Welcome! |
|
Emerging |
| 38 |
sinfallas/opendevin-docker
Run OpenDevin inside Docker |
|
Emerging |
| 39 |
Jewelzufo/granitepi-4-nano
Run IBM Granite 4.0 locally on Raspberry Pi 5 with Ollama.This is a... |
|
Emerging |
| 40 |
Scottcjn/llama-cpp-tigerleopard
WORLD FIRST: llama.cpp for Mac OS X Tiger & Leopard on PowerPC G4/G5 |
|
Emerging |
| 41 |
aws-samples/sample-ollama-server
Ollama on GPU EC2 instance with Open WebUI web interface and Bedrock access |
|
Emerging |
| 42 |
AbhinaavRamesh/ollama-local-serve
Local LLM infrastructure for distributed AI applications. Serve... |
|
Emerging |
| 43 |
ai-action/setup-ollama
🦙 Set up GitHub Actions with Ollama CLI |
|
Emerging |
| 44 |
teremterem/litellm-server-boilerplate
A lightweight LiteLLM server boilerplate pre-configured with uv and Docker... |
|
Emerging |
| 45 |
EvilFreelancer/docker-llama.cpp-rpc
Данный проект основан на llama.cpp и компилирует только RPC-сервер, а так же... |
|
Emerging |
| 46 |
heyvaldemar/ollama-traefik-letsencrypt-docker-compose
Ollama with Let's Encrypt Using Docker Compose |
|
Emerging |
| 47 |
rgryta/LLM-WSL2-Docker
One-click install for WizardLM-13B-Uncensored with oobabooga webui |
|
Emerging |
| 48 |
sasha0552/ToriLinux
Linux LiveCD for offline AI training and inference. |
|
Emerging |
| 49 |
mitja/llamatunnel
Publish local LLMs and LLM apps on the internet. |
|
Emerging |
| 50 |
raketenkater/llm-server
Smart launcher for llama.cpp / ik_llama.cpp — auto-detects GPUs, optimizes... |
|
Experimental |
| 51 |
azer/llmcat
Prepare files and directories for LLM consumption |
|
Experimental |
| 52 |
crowdllama/crowdllama
CrowdLlama is a distributed system that leverages the open-source Ollama... |
|
Experimental |
| 53 |
asreview/asreview-server-stack
Docker compose for setting up ASReview server with authentication |
|
Experimental |
| 54 |
linonetwo/MOSS-DockerFile
用于在 Docker 里运行复旦的 MOSS 语言模型,使用 GradIO 提供 WebUI。 |
|
Experimental |
| 55 |
toku345/dgx-llm-serve
Docker Compose configs for running LLM inference on DGX Spark (TensorRT-LLM... |
|
Experimental |
| 56 |
Scottcjn/power8-projects
POWER8 Projects - Ubuntu 22.04 build, PSE LLM, Darwin cross-compile |
|
Experimental |
| 57 |
m1ns09/Llama
🌐 Run GGUF models directly in your web browser using JavaScript and... |
|
Experimental |
| 58 |
GURPREETKAURJETHRA/Ollama-UseCases
This repo brings numerous use cases from the Open Source Ollama |
|
Experimental |
| 59 |
soulteary/docker-yi-runtime
零一万物(34B)的本地运行环境。 |
|
Experimental |
| 60 |
alex0dd/llm-app-microservices-template
Template for building microservice-based apps with a frontend, backend, LLM... |
|
Experimental |
| 61 |
ivangabriele-archives/docker-llm
Pre-loaded LLMs served as an OpenAI-Compatible API via Docker images. |
|
Experimental |
| 62 |
codygreen/llm_api_server
Lab to demonstrate how to apply an API to an AI model and secure it. |
|
Experimental |
| 63 |
g1ibby/homellm
A simple Docker Compose boilerplate for deploying Open WebUI and LiteLLM... |
|
Experimental |
| 64 |
LianHe-BI/Blackwell-optimized-llama.cpp-Docker-image
Blackwell-optimized llama.cpp Docker image – works on all NVIDIA GPUs, but... |
|
Experimental |
| 65 |
muhac/llm-actions
Run LLMs for inference in GitHub Actions - add to your workflow! |
|
Experimental |
| 66 |
AnLaVN/AL-Library
Java utility library, contain many feature, support to Large Language Model... |
|
Experimental |
| 67 |
DataJourneyHQ/list-github-models
GitHub action to track GitHub Models |
|
Experimental |
| 68 |
ivangabriele-archives/docker-functionary
Ready-to-deploy Docker image for Functionary LLM served as an OpenAI-Compatible API. |
|
Experimental |
| 69 |
Malax/buildpack-ollama
Cloud Native Buildpack that builds an OCI image with Ollama and a large... |
|
Experimental |
| 70 |
wizzard0/llama2.ts
Llama2 inference in one TypeScript file |
|
Experimental |
| 71 |
OutofAi/ChitChat
Modal LLM LLama.cpp based model deployment as part of series of Model as a... |
|
Experimental |
| 72 |
JimKw1kX/LLM-C2-Server
An AI C2 Server |
|
Experimental |
| 73 |
mordang7/LlamaForge
The Ultimate Command Center for Local LLMs. A professional-grade GUI for... |
|
Experimental |
| 74 |
rookiemann/vllm-windows-build
Native Windows build patches for vLLM v0.14.1 — MSVC 2022 + CUDA 12.6, 26... |
|
Experimental |
| 75 |
arseniy0924/rpc_manager
Web UI for orchestrating distributed llama.cpp RPC GPU clusters with auto... |
|
Experimental |
| 76 |
nyo16/llama_cpp_ex
Elixir bindings for llama.cpp — run LLMs locally with Metal, CUDA, Vulkan,... |
|
Experimental |
| 77 |
futursolo/pai
Collection of AI Containers - Prebuilt and Ready-to-Use |
|
Experimental |
| 78 |
ggalancs/hfl
CLI + API server to download, manage, and run 500K+ HuggingFace models... |
|
Experimental |
| 79 |
mysticrenji/ollama-on-kubernetes
An attempt to run ollama on kuberenetes |
|
Experimental |
| 80 |
micbi-dt/lmstudio-docker
run LMStudio within a Docker container |
|
Experimental |
| 81 |
alasgarovs/openserv
OpenServ is a simple Bash-based CLI tool for managing LLMs in llama.cpp server. |
|
Experimental |
| 82 |
SergiuDeveloper/distributed-llama.cpp
Distributed LLM inference across multiple machines. A central server routes... |
|
Experimental |
| 83 |
SuppieRK/local-ai-lab
Offline-capable, open-source AI home lab notes: practical setups, configs,... |
|
Experimental |
| 84 |
onidahabitual85/llm-server
Launch and optimize llama.cpp servers automatically across Linux, macOS, and... |
|
Experimental |
| 85 |
Skyluker4/llama-runpod
Docker image to run llama.cpp on runpod.io automatically |
|
Experimental |
| 86 |
Pavloffm/remote-llm-server
Run Ollama in Docker. Share local LLMs across your network. GPU-accelerated. |
|
Experimental |
| 87 |
qnianjinri-del/local-llm-recommender
一键识别电脑硬件,推荐最新适配的开源大模型,并支持一键部署。 |
|
Experimental |
| 88 |
rjxby/llama-runtime
`llama-runtime` is a high-performance inference server designed for local... |
|
Experimental |
| 89 |
gsavla6-hue/java-llm-integration
Comprehensive Java LLM integration library supporting OpenAI, Anthropic and... |
|
Experimental |
| 90 |
Daaboulex/lmstudio-nix
LM Studio packaged for NixOS — local LLM inference desktop app and server |
|
Experimental |
| 91 |
llmjava/hf_text_generation
Hugging Face Text Generation API client for Java |
|
Experimental |
| 92 |
EricApgar/llm-server
Host an LLM and make it accessible on a network via API. |
|
Experimental |
| 93 |
openradx/llm_api_server_mock
This is a simple fastapi based server mock that implements the OpenAI API. |
|
Experimental |
| 94 |
clixgvvv/AndroidLLMServerScript
📲 Create a local LLM server on Android using Python and llama.cpp for easy... |
|
Experimental |
| 95 |
tdiprima/ollama-orchestrator
Self-hosted AI automation: manage Ollama models, deploy Open WebUI in... |
|
Experimental |
| 96 |
MooNyeu/granitepi-4-nano
🔒 Run a large language model locally on your Raspberry Pi 5 with IBM Granite... |
|
Experimental |
| 97 |
sithukyaw007/local-ai-workload
Docker-first, local-first AI workload toolkit for macOS Apple Silicon using... |
|
Experimental |
| 98 |
Logicish/p-lanes
A modular wrapper for llama.cpp focused on home-lab scaled hardware,... |
|
Experimental |
| 99 |
abdulazizalmalki-gh/local-ai
A simple, self-hosted stack for running AI models locally using llama.cpp... |
|
Experimental |
| 100 |
byang37/llama-runner
A lightweight desktop GUI for llama-server — multi-model routing, per-model... |
|
Experimental |
| 101 |
dmeldrum6/Llama-Forge
Open source llama.cpp wrapper with server and client |
|
Experimental |
| 102 |
gperdrizet/llms-devcontainer
Containerized development environment for LLM based projects |
|
Experimental |
| 103 |
AiratTop/ollama-self-hosted
A simple Docker Compose setup to self-host Ollama and Open WebUI. Run your... |
|
Experimental |
| 104 |
b-data/mojo-docker-stack
(GPU accelerated) Multi-arch (linux/amd64, linux/arm64/v8) MAX/Mojo docker... |
|
Experimental |
| 105 |
mdaconta/xlm-eco-api
Cross Language Model (LLM/SLM/etc.) Ecosystem API (xlm-eco-api) |
|
Experimental |
| 106 |
hoonywise/minerva
A private, GPU-accelerated AI stack with Ollama, LangChain, Stable... |
|
Experimental |
| 107 |
zyoung11/lmgo
Windows system tray for llama.cpp + ROCm. Optimized for AMD RYZEN AI MAX+... |
|
Experimental |
| 108 |
rookiemann/llama-cpp-python-py314-cuda131-wheel
GPU-accelerated llama-cpp-python 0.3.16 wheel for Python 3.14 (CUDA 13.1, Windows) |
|
Experimental |
| 109 |
qianniuspace/movie-detectives-server
骆驼电影侦探社(服务端) |
|
Experimental |
| 110 |
mo-arvan/local-llm
docker compose configuration file for running Llama-2 or any other language... |
|
Experimental |
| 111 |
yokingma/deepseek-vllm
Docker&vLLM官方镜像部署DeepSeek模型,在生产环境中提供类OpenAI接口服务。 |
|
Experimental |
| 112 |
kryoz/llama-strix-halo
llama.cpp setup on dedicated AMD Strix Halo machine |
|
Experimental |
| 113 |
ai-action/ollama-github-action-demo
🦙 Demos of large language models (LLMs) with Ollama in GitHub Actions. |
|
Experimental |
| 114 |
FlorinAndrei/local-inference-docs
Run generative AI locally, on your hardware, for coding and other purposes |
|
Experimental |
| 115 |
ThomasVitale/llm-images
Catalog of OCI images for popular open-source or open Large Language Models. |
|
Experimental |
| 116 |
stlin256/llama-remote
A web-based remote control panel for managing llama.cpp instances. Monitor... |
|
Experimental |
| 117 |
abhiFSD/llama.cpp-Monitor-Dashboard
⚡ Real-time monitoring dashboard for llama.cpp server — single HTML file,... |
|
Experimental |
| 118 |
somya-droid/Pirate-LLM-Server
Run local LLM servers on iPhone with OpenAI-compatible API, Metal GPU... |
|
Experimental |
| 119 |
ebowwa-archive/LLM_telecenter
A fastapi wrapper of babca / python-gsmmodem for a waveshare sim7600x. Not... |
|
Experimental |
| 120 |
sebicom/llamacpp4j
Java wrapper for llama.cpp |
|
Experimental |
| 121 |
nishantapatil3/litellm-compose
Docker Compose setup for LiteLLM proxy server with PostgreSQL and Prometheus... |
|
Experimental |
| 122 |
beeracs/Llama
Run Llama models in your web browser using JavaScript and WebAssembly.... |
|
Experimental |
| 123 |
llmjava/llm4j
One API to access Large Language Models in Java |
|
Experimental |
| 124 |
andrewginns/LocalLLM
Configurations for a locally hosted LLM and applications leveraging it |
|
Experimental |
| 125 |
VityazevEgor/LLMapi4free
LLMapi4free provides a unified API for free access to various large Language... |
|
Experimental |
| 126 |
buckyinsfo/homelab-ai-stack
Self-hosted AI + GPU server homelab — local LLM inference, vector search,... |
|
Experimental |
| 127 |
Weebaay/local-ai-homelab
Déploiement d'un serveur IA local sur VM Ubuntu Server 24.04 avec Ollama et... |
|
Experimental |
| 128 |
mendhak/local-llm-workspace
Private, secure, containerized LLM environment for chat and coding. Using... |
|
Experimental |
| 129 |
Riju007/dev-knowledge-vault
🧠 My second brain — hands-on engineering notes on Docker, AI, Python and beyond |
|
Experimental |
| 130 |
chaserbot/chaseworkslab-llm
Self-hosted LLM stack (Ollama, Open WebUI, etc.) for the homelab |
|
Experimental |
| 131 |
cyberguard-ai/local-llm-server
A containerized, offline-capable LLM API powered by Ollama. Automatically... |
|
Experimental |
| 132 |
nishant-sethi/python-ai-extension-server
Python Server to use local LLMs |
|
Experimental |
| 133 |
thkox/home-ai-server
Home AI Server provides the backend infrastructure for the Home AI system.... |
|
Experimental |
| 134 |
sinfallas/llm-local-loader-docker
docker compose to load ollama, flowise, langfuse, open-web-ui |
|
Experimental |
| 135 |
57Ajay/model-runner
A simple model runner using llama.cpp and huggingface |
|
Experimental |
| 136 |
gustavostz/Local-AI-Open-Orca-For-Dummies
Local AI Open Orca For Dummies is a user-friendly guide to running Large... |
|
Experimental |
| 137 |
FarzamMohammadi/self-hosted-ai-stack
Blog resources for building a self-hosted AI infrastructure. Contains all... |
|
Experimental |
| 138 |
merlijn/scala-llm-api
Basic OpenAI client for Scala |
|
Experimental |
| 139 |
wronai/docker-platform
Enterprise-grade secure media storage with AI analysis, role-based access,... |
|
Experimental |
| 140 |
AntonSHBK/llm_service
A FastAPI-based microservice for interacting with LLM (OpenAI API) with... |
|
Experimental |
| 141 |
ai-action/ai-inference-demo
AI Inference in GitHub Actions demo |
|
Experimental |
| 142 |
yeeking/llamacpp-minimal-example
Minimal example of using llama cpp as library from cpp |
|
Experimental |
| 143 |
aayes89/JavaRNN-LLM
An RNN written in pure Java to compete with Transformers |
|
Experimental |
| 144 |
theomart/llm-based-api-template
🐣 A template to deploy an LLM based API to Cloud Run, using FastAPI, Docker... |
|
Experimental |
| 145 |
desdeux/llama2odin
Llama2.C port in Odin |
|
Experimental |
| 146 |
Doculoom/doculoom-server
LLM backed API server |
|
Experimental |
| 147 |
abhishekrana/llm-service
RESTful service with LLMs (Large Language Models) running locally |
|
Experimental |
| 148 |
turtleio/turtle
🐰 shoulda been an app - 🐢 |
|
Experimental |
| 149 |
aryansingla45/flask-llm-ci-cd
The app allows users to upload files, which are stored in a dedicated... |
|
Experimental |
| 150 |
MrTechyWorker/SmartLLM-Server
Implementing a robust client-server architecture from scratch, designed to... |
|
Experimental |