Local LLM Deployment Transformer Models
Tools and resources for running, hosting, and serving open-source LLMs locally or on private infrastructure without cloud dependencies. Includes deployment platforms, free API gateways, optimization guides, and access control for self-hosted models. Does NOT include model training, fine-tuning frameworks, or cloud-based LLM services.
There are 245 local llm deployment models tracked. 4 score above 70 (verified tier). The highest-rated is withcatai/node-llama-cpp at 79/100 with 1,942 stars and 4,219,393 monthly downloads. 9 of the top 10 are actively maintained.
Get all 245 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=local-llm-deployment&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp.... |
|
Verified |
| 2 |
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models |
|
Verified |
| 3 |
bentoml/OpenLLM
Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible... |
|
Verified |
| 4 |
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others.... |
|
Verified |
| 5 |
SciSharp/LLamaSharp
A C#/.NET library to run LLM (π¦LLaMA/LLaVA) on your local device efficiently. |
|
Established |
| 6 |
zhudotexe/kani
kani (γ«γ) is a highly hackable microframework for tool-calling language... |
|
Established |
| 7 |
mostlygeek/llama-swap
Reliable model swapping for any local OpenAI/Anthropic compatible server -... |
|
Established |
| 8 |
Michael-A-Kuykendall/shimmy
β‘ Python-free Rust inference server β OpenAI-API compatible. GGUF +... |
|
Established |
| 9 |
UbiquitousLearning/mllm
Fast Multimodal LLM on Mobile Devices |
|
Established |
| 10 |
kaito-project/aikit
ποΈ Fine-tune, build, and deploy open-source LLMs easily! |
|
Established |
| 11 |
mybigday/llama.rn
React Native binding of llama.cpp |
|
Established |
| 12 |
cheahjs/free-llm-api-resources
A list of free LLM inference resources accessible via API. |
|
Established |
| 13 |
sgl-project/ome
Open Model Engine (OME) β Kubernetes operator for LLM serving, GPU... |
|
Established |
| 14 |
floneum/floneum
Instant, controllable, local pre-trained AI models in Rust |
|
Established |
| 15 |
Mobile-Artificial-Intelligence/llama_sdk
lcpp is a dart implementation of llama.cpp used by the mobile artificial... |
|
Established |
| 16 |
tattn/LocalLLMClient
Swift package to run local LLMs on iOS, macOS, Linux |
|
Established |
| 17 |
Strvm/meta-ai-api
Llama 3 API 70B & 405B (MetaAI Reverse Engineered) |
|
Established |
| 18 |
mukel/llama3.java
Practical Llama 3 inference in Java |
|
Emerging |
| 19 |
guinmoon/LLMFarm
llama and other large language models on iOS and MacOS offline using GGML library. |
|
Emerging |
| 20 |
mirpo/fastapi-gen
Build LLM-enabled FastAPI applications without build configuration. |
|
Emerging |
| 21 |
belladoreai/llama3-tokenizer-js
JS tokenizer for LLaMA 3 and LLaMA 3.1 |
|
Emerging |
| 22 |
nekomeowww/ollama-operator
π’ Yet another operator for running large language models on Kubernetes with... |
|
Emerging |
| 23 |
guinmoon/llmfarm_core.swift
Swift library to work with llama and other large language models. |
|
Emerging |
| 24 |
tairov/llama2.mojo
Inference Llama 2 in one file of pure π₯ |
|
Emerging |
| 25 |
mfoud444/ollamafreeapi
OllamaFreeAPI: Free Distributed API for Ollama LLMs Public gateway to our... |
|
Emerging |
| 26 |
tjake/Jlama
Jlama is a modern LLM inference engine for Java |
|
Emerging |
| 27 |
BeRo1985/pasllm
PasLLM - LLM inference engine in Object Pascal (synced from my private work... |
|
Emerging |
| 28 |
local-ai-zone/local-ai-zone.github.io
Discover the Best AI Models for Your PC |
|
Emerging |
| 29 |
yoshoku/llama_cpp.rb
llama_cpp.rb provides Ruby bindings for llama.cpp |
|
Emerging |
| 30 |
camenduru/text-generation-webui-colab
A colab gradio web UI for running Large Language Models |
|
Emerging |
| 31 |
sammcj/ingest
Parse files (e.g. code repos) and websites to clipboard or a file for... |
|
Emerging |
| 32 |
LM-Kit/lm-kit-net-samples
.NET samples for LM-Kit.NET |
|
Emerging |
| 33 |
nova-land/gbnf-compiler
Plug n Play GBNF Compiler for llama.cpp |
|
Emerging |
| 34 |
ngxson/wllama
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference |
|
Emerging |
| 35 |
fboulnois/llama-cpp-docker
Run llama.cpp in a GPU accelerated Docker container |
|
Emerging |
| 36 |
jmont-dev/ollama-hpp
Modern, Header-only C++ bindings for the Ollama API. |
|
Emerging |
| 37 |
hybridgroup/yzma
Go with your own intelligence - Go applications that directly integrate... |
|
Emerging |
| 38 |
soulteary/docker-llama2-chat
Play LLaMA2 (official / δΈζη / INT4 / llama2.cpp) Together! ONLY 3 STEPS! (... |
|
Emerging |
| 39 |
withcaer/curtana
Simplified zero-cost wrapper over llama.cpp powered by the lama-cpp-2 Crate. |
|
Emerging |
| 40 |
Archimedes1618/Madlab
Madlab is an advanced AI development studio designed to streamline the... |
|
Emerging |
| 41 |
cocktailpeanut/dalai
The simplest way to run LLaMA on your local machine |
|
Emerging |
| 42 |
sobelio/llm-chain
`llm-chain` is a powerful rust crate for building chains in large language... |
|
Emerging |
| 43 |
donderom/llm4s
Scala 3 bindings for llama.cpp π¦ |
|
Emerging |
| 44 |
absadiki/pyllamacpp
Python bindings for llama.cpp |
|
Emerging |
| 45 |
iaalm/llama-api-server
A OpenAI API compatible REST server for llama. |
|
Emerging |
| 46 |
mdrokz/rust-llama.cpp
LLama.cpp rust bindings |
|
Emerging |
| 47 |
loong64/llama.cpp
LLM inference in C/C++ |
|
Emerging |
| 48 |
openjlc/riscv64-library
Some of the libraries (docs) on the RISCV64 architecture are easy for users... |
|
Emerging |
| 49 |
gitctrlx/llama.go
Llama from scratch in Go. |
|
Emerging |
| 50 |
nerve-sparks/iris_android
IRIS is an android app for interfacing with GGUF / llama.cpp models locally. |
|
Emerging |
| 51 |
nuhmanpk/quick-llama
Run Ollama models on Google Colab |
|
Emerging |
| 52 |
LLukas22/llm-rs-python
Unofficial python bindings for the rust llm library. πβ€οΈπ¦ |
|
Emerging |
| 53 |
gotzmann/llama.go
llama.go is like llama.cpp in pure Golang! |
|
Emerging |
| 54 |
diogok/llama.cpp.zig
A build.zig for llama.cpp |
|
Emerging |
| 55 |
loong64/ollama
Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 2, and other... |
|
Emerging |
| 56 |
KolosalAI/kolosal-server
Kolosal AI is an OpenSource and Lightweight alternative to Ollama to run... |
|
Emerging |
| 57 |
mybigday/llama.node
Node.js binding of llama.cpp |
|
Emerging |
| 58 |
phronmophobic/llama.clj
Run LLMs locally. A clojure wrapper for llama.cpp. |
|
Emerging |
| 59 |
developer239/llama.cpp-ts
llama.cpp π¦ LLM inference in TypeScript |
|
Emerging |
| 60 |
KolosalAI/kolosal-cli
Super lightweight Ollama + Qwen Code alternative to run Llama 3.3,... |
|
Emerging |
| 61 |
fardjad/node-llmatic
Use self-hosted LLMs with an OpenAI compatible API |
|
Emerging |
| 62 |
eugenehp/bitnet-cpp-rs
Rust bindings for bitnet.cpp based on llama-cpp-4 |
|
Emerging |
| 63 |
KolosalAI/Kolosal
Kolosal AI is an OpenSource and Lightweight alternative to LM Studio to run... |
|
Emerging |
| 64 |
BodhiSearch/BodhiApp
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs |
|
Emerging |
| 65 |
trrahul/llama2.cs
Inference Llama 2 in one file of pure C# |
|
Emerging |
| 66 |
iverly/llamafile-docker
Distribute and run llamafile/LLMs with a single docker image. |
|
Emerging |
| 67 |
cgbur/llama2.zig
Inference Llama 2 in one file of pure Zig |
|
Emerging |
| 68 |
dirmacs/lancor
A Rust client library for llama.cpp's OpenAI-compatible API server |
|
Emerging |
| 69 |
hpretila/llama.net
.NET wrapper for LLaMA.cpp for LLaMA language model inference on CPU. π¦ |
|
Emerging |
| 70 |
belladoreai/llama-tokenizer-js
JS tokenizer for LLaMA 1 and 2 |
|
Emerging |
| 71 |
anthonyfoust/ai-stack-homelab
Complete AI automation stack optimized for Mac Mini M4, but can work in... |
|
Emerging |
| 72 |
mdegans/drama_llama
Yet another `llama.cpp` Rust wrapper |
|
Emerging |
| 73 |
amin-tehrani/ollama-colab
Serve Ollama LLMs on Google Colab (free plan) using Ngrok |
|
Emerging |
| 74 |
jaco-bro/MLX.zig
MLX.zig: Phi-4, Llama 3.2, and Whisper in Zig |
|
Emerging |
| 75 |
Kagamma/llama-pas
Free Pascal bindings for llama.cpp |
|
Emerging |
| 76 |
Thrasher-Software/sigil
A local-first LLM development studio. Build, test, and customize inference... |
|
Emerging |
| 77 |
Agora-Lab-AI/Atom
a suite of finetuned LLMs for atomically precise function calling π§ͺ |
|
Emerging |
| 78 |
SeungyounShin/Llama2-Code-Interpreter
Make Llama2 use Code Execution, Debug, Save Code, Reuse it, Access to Internet |
|
Emerging |
| 79 |
FlatlinerDOA/PerceptivePyro
Run and train Transformer based Large Language Models (LLMS) natively in... |
|
Emerging |
| 80 |
adalkiran/llama-nuts-and-bolts
A holistic way of understanding how Llama and its components run in... |
|
Emerging |
| 81 |
openshieldai/openshield
OpenShield is a new generation security layer for AI models |
|
Emerging |
| 82 |
K024/llm-sharp
Language models in C# |
|
Emerging |
| 83 |
abhisheknair10/llama3.cu
Lightweight Llama 3 8B Inference Engine in CUDA C |
|
Experimental |
| 84 |
dravenk/ollama-zig
Ollama Zig library |
|
Experimental |
| 85 |
trzy/llava-cpp-server
LLaVA server (llama.cpp). |
|
Experimental |
| 86 |
dev-sufyaan/Nexlify
Unified API platform for free access to enterprise-grade AI models from... |
|
Experimental |
| 87 |
Aloereed/llama.cpp-server-ohos
Llama.cpp server for OpenHarmony |
|
Experimental |
| 88 |
c0sogi/llama-api
An OpenAI-like LLaMA inference API |
|
Experimental |
| 89 |
sashazykov/json-repair-rb
A simple Ruby gem designed to repair broken JSON strings |
|
Experimental |
| 90 |
lrusso/llama3pure
Three inference engines for Llama 3: pure C for desktop systems, pure... |
|
Experimental |
| 91 |
nikolaydubina/llama2.go
LLaMA-2 in native Go |
|
Experimental |
| 92 |
hoof-ai/hoof
"Just hoof it!" - A spotlight like interface to Ollama |
|
Experimental |
| 93 |
saddam213/LLamaStack
ASP.NET Core Web, WebApi & WPF implementations for LLama.cpp & LLamaSharp |
|
Experimental |
| 94 |
leftmove/cria
Run LLMs locally with as little friction as possible. |
|
Experimental |
| 95 |
andreiramani/jadi4llamacpp
Just another drop in for llama.cpp |
|
Experimental |
| 96 |
OneInterface/realtime-bakllava
llama.cpp with BakLLaVA model describes what does it see |
|
Experimental |
| 97 |
fermyon/ai-examples
A collection of serverless apps that show how Fermyon's Serverless AI... |
|
Experimental |
| 98 |
chelsea0x3b/llama-dfdx
LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed! |
|
Experimental |
| 99 |
yfedoseev/llmkit
Production-grade LLM client - Rust, Python, TypeScript. 100+ providers,... |
|
Experimental |
| 100 |
moritztng/fltr
Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B. |
|
Experimental |
| 101 |
maifeeulasad/LocalLLaMA
π LocalLLaMA Archive β Community-powered static archive for r/LocalLLaMA |
|
Experimental |
| 102 |
5aharsh/collama
Run Ollama LLM models in Google Colab for free |
|
Experimental |
| 103 |
daskol/llama.py
Python bindings to llama.cpp |
|
Experimental |
| 104 |
zTgx/llmweb-rs
Webpage to structured data in Rust & LLM |
|
Experimental |
| 105 |
zerob13/modelinfo-cli
A CLI to query AI model capabilities, context limits, and pricing from... |
|
Experimental |
| 106 |
AlenVelocity/langchain-llama
Run LLAMA LLMs in Node with Langchain |
|
Experimental |
| 107 |
zatevakhin/obsidian-local-llm
Obsidian Local LLM is a plugin for Obsidian that provides access to a... |
|
Experimental |
| 108 |
Uralstech/vid-orca
Deploy LLaMA-2 Chat on Google Cloud. |
|
Experimental |
| 109 |
cgjosephlee/ollama-save-load
Save and load ollama models just like operating docker images. |
|
Experimental |
| 110 |
benct/kotlin-cheat-sheet
:star: Kotlin <3 Cheat Sheet, Collection Extension Functions and General Examples |
|
Experimental |
| 111 |
johnsutor/llama-jarvis
Turn any LLM into Jarvis |
|
Experimental |
| 112 |
didier-durand/llms-in-clouds
Experiments with LLMs in clouds (powered by SGLang) |
|
Experimental |
| 113 |
kassane/ollama-d
D bindings for the Ollama API |
|
Experimental |
| 114 |
codewithdark-git/llama-3-Hackathon
LLaMA Genius is an AI-powered research assistant designed to help users... |
|
Experimental |
| 115 |
khiwniti/kaggle-llm-api
π€ Comprehensive solution for running Ollama/vLLM API servers in Kaggle... |
|
Experimental |
| 116 |
AI-Robotic-Labs/Self-Sovereign-AI-SDK
SDK for Self Sovereign AI |
|
Experimental |
| 117 |
BerkeliumLabs/Berkelium-labs
Your personal AI Lab, accessible everywhere! Explore, experiment, and... |
|
Experimental |
| 118 |
alvion427/PerroPastor
Run Llama based LLMs in Unity entirely in compute shaders with no dependencies |
|
Experimental |
| 119 |
rabilrbl/llamafile-builder
A simple github actions script to build a llamafile and uploads to huggingface |
|
Experimental |
| 120 |
cvedix/omnisdk
On-device AI deloper platform |
|
Experimental |
| 121 |
excorsistvoid/Neuro-Bridge
π Enable seamless hardware access on Android with Neuro-Bridge, a... |
|
Experimental |
| 122 |
RahulSChand/llama2.c-for-dummies
Step by step explanation/tutorial of llama2.c |
|
Experimental |
| 123 |
avatsaev/av-local-llm-api
Allows to easily run local REST API with a custom LLM, running locally or... |
|
Experimental |
| 124 |
makllama/makllama
MaK(Mac+Kubernetes)llama - Running and orchestrating large language models... |
|
Experimental |
| 125 |
Brazilian-willametteriver232/llama.swift
π Access llama.cpp easily in your Swift projects, leveraging precompiled... |
|
Experimental |
| 126 |
ksylvest/omniai-llama
An implementation of the OmniAI interface for Llama. |
|
Experimental |
| 127 |
frinknet/gelli
Containerized LLM for any use-case big or small |
|
Experimental |
| 128 |
RichardHam-co-uk/ProjectLodestar
AI development environment with 90% cost savings. Routes between 8 LLM... |
|
Experimental |
| 129 |
PCfVW/plip-rs
Mechanistic interpretability toolkit for code LLMs, in Rust. Analysis of... |
|
Experimental |
| 130 |
seanpm2001/DALL-E_LLaMA_Docs
π€οΈπ¦οΈπ§ οΈποΈ The official documentation source repository for DALL-E LLaMA, a... |
|
Experimental |
| 131 |
kurnevsky/llama-cpp.el
A client for llama-cpp server |
|
Experimental |
| 132 |
seanpm2001/DALL-E_LLaMA
π€οΈπ¦οΈπ§ οΈ DALL-E LLaMA is a combination of DALL-E and LLaMA (Large Language... |
|
Experimental |
| 133 |
tunib-ai/joker
AI model designed to test the effectiveness in handling external ethical attacks. |
|
Experimental |
| 134 |
luiscavallcante859/collectiv-ai-sdk
π Build and integrate with the CollectiVAI Router using official SDKs for... |
|
Experimental |
| 135 |
veerapatel/llm.nexus
π Streamline integration with various LLM providers using LLM.Nexus, a .NET... |
|
Experimental |
| 136 |
UgurkanTech/ArchNetAI
ArchNetAI is a Python library that leverages the Ollama API for generating... |
|
Experimental |
| 137 |
Adriankhl/godot-llm-template
Godot LLM Template/Demo |
|
Experimental |
| 138 |
estrify/ProjectLodestar
π Optimize AI development with Lodestar by smartly routing between free... |
|
Experimental |
| 139 |
nininau/awesome-llm-services
π Discover 106+ open-source LLM services and tools for AI, ideal for local... |
|
Experimental |
| 140 |
ahmedmagood/cpu-slm
π₯οΈ Explore CPU-SLM, a Rust-based SLM/LLM project that runs on CPU, offering... |
|
Experimental |
| 141 |
whyisitworking/llama-bro
On-device LLM inference SDK for Android, powered by llama.cpp. Run GGUF... |
|
Experimental |
| 142 |
ferranpons/Llamatik-Server
Remote inference backend implementing the same API as the Llamatik library... |
|
Experimental |
| 143 |
NeoZel/huatuo
π Enhance your cloud-native observability with HUATUO, using eBPF for deep... |
|
Experimental |
| 144 |
ns408/local-ai-setup
Run modern AI models on older laptops - optimized for 2nd-gen Intel hardware |
|
Experimental |
| 145 |
qxoticai/qxotic
AI engine for the JVM |
|
Experimental |
| 146 |
wk-y/rama-swap
ramalama-based model swapping server |
|
Experimental |
| 147 |
blackboxprogramming/ai-chain
AI Chain β Distributed multi-node LLM inference with automatic failover.... |
|
Experimental |
| 148 |
fuglede/llama.ttf
A font for writing tiny stories |
|
Experimental |
| 149 |
ariannamethod/yent.yo
diffusion AI with a bad character |
|
Experimental |
| 150 |
hurui200320/llama-cpp-kt
The Kotlin wrapper of llama.cpp, powered by JNA |
|
Experimental |
| 151 |
Root1V/axonium-sdk
A production-grade Python SDK for llama-server that streamlines... |
|
Experimental |
| 152 |
anglerfishlyy/llm-watch-grafana
AI observability Grafana plugin tracking real-time LLM metrics β latency,... |
|
Experimental |
| 153 |
fbaldassarri/llama-cpp-container
Docker image to deploy a llama-cpp container with conda-ready environments |
|
Experimental |
| 154 |
bkataru/chatllm.zig
Zig wrapper for chatllm.cpp - LLM inference with 70+ model architectures |
|
Experimental |
| 155 |
Stoksweet/modlable
A platform for building, training and running inference on TensorflowJS... |
|
Experimental |
| 156 |
haormj/llama2.go
Inference Llama 2 in one file of pure go |
|
Experimental |
| 157 |
LastBotInc/llama2j
Pure Java Llama2 inference with optional multi-GPU CUDA implementation |
|
Experimental |
| 158 |
chromejaw/free-llm-api
A list of free LLM inference resources accessible via API. |
|
Experimental |
| 159 |
lwch/llama2.go
Port of Facebook's LLaMA 2 model in pure go and use little memory |
|
Experimental |
| 160 |
leaxer-ai/leaxer-llama
Pre-built llama.cpp binaries for Leaxer |
|
Experimental |
| 161 |
Komdosh/kLLaMa-jvm
Simple example of using llama.cpp with kotlin (JVM) |
|
Experimental |
| 162 |
revengerrr/LedgerCOBOL
A COBOL banking system with AI integration. Built to learn how legacy code... |
|
Experimental |
| 163 |
EZForever/llama.cpp-static
Static builds of llama.cpp (Currently only amd64 server builds are available) |
|
Experimental |
| 164 |
tokenrouter/tokenrouter-python
Official Python SDK for TokenRouter - an intelligent LLM routing service... |
|
Experimental |
| 165 |
invergent-ai/surogate-website
Website for surogate.ai |
|
Experimental |
| 166 |
instavm/llm-token-visualizer
See How Big Exactly A 128k Token Text Is |
|
Experimental |
| 167 |
mhajder/llama.cpp-updater
A shell script to automatically update or build llama.cpp with optimal GPU... |
|
Experimental |
| 168 |
Andrew2077/Alpaca
Simple Q/A app, where i created a UI for alpaca (fine tuned LLAMA) model... |
|
Experimental |
| 169 |
LlamaGenAI/llamagenai-openapi
LlamaGen.Ai REST API, LlamaGen is AI Comic Factory - Generate Comics with... |
|
Experimental |
| 170 |
sc0v0ne/udemy_course_mastering_ollama_build_private_local_llm_apps_with_python
Udemy Course Mastering Ollama Build Private Local LLM Apps with Python |
|
Experimental |
| 171 |
lenticularis39/llama2.inferno
Inference Llama 2 in one file of pure Limbo |
|
Experimental |
| 172 |
waqasm86/Ubuntu-Cuda-Llama.cpp-Executable
Pre-built llama.cpp CUDA binary for Ubuntu 22.04. No compilation required -... |
|
Experimental |
| 173 |
Gaolingx/llama.cpp-Launcher
run llama.cpp quickly and conveniently. |
|
Experimental |
| 174 |
lennor-tan/openrouter-free-model
π Explore and manage free models on OpenRouter effortlessly with our web... |
|
Experimental |
| 175 |
entelecheia/llama-factory-container
Container for LLaMA-Factory |
|
Experimental |
| 176 |
GP-Silah/silah-ai
Powering Silah's smart features! |
|
Experimental |
| 177 |
KolosalAI/kolosal-desktop
Kolosal AI is an OpenSource and Lightweight alternative to LM Studio to run... |
|
Experimental |
| 178 |
mrtrizer/UnityLlamaCpp
Llama.cpp in Unity, straightforward and clean |
|
Experimental |
| 179 |
austinweis/alpaca.cpp-gui
GUI for GGML Alpaca models |
|
Experimental |
| 180 |
ChristianHohlfeld/ollama-local-docker
Ollama Local Docker - A simple Docker-based setup for running Ollama's API... |
|
Experimental |
| 181 |
harpertoken/memoraxx
LLaMA-style models with memory persistence. |
|
Experimental |
| 182 |
sak96/rust_llama_app
Chat bot (llama) written in rust using Yew and Tauri. |
|
Experimental |
| 183 |
jihadkhawaja/Llama.Grammar
GBNF converter for llama.cpp Grammar directly from C# types |
|
Experimental |
| 184 |
juansalnac/API-mega-list
π Discover a comprehensive collection of APIs to enhance your projects and... |
|
Experimental |
| 185 |
nathanborror/swift-llama
A Swift client library for interacting with Meta's Llama API. |
|
Experimental |
| 186 |
secret-ai-labs/awesome-local-llm
Your complete guide to running powerful AI models locally in 2025. Covers... |
|
Experimental |
| 187 |
nerdsupremacist/LlamaLang
Repository for the Llama Programming Language. Work In Progress |
|
Experimental |
| 188 |
seehiong/micronaut-llama3
A high-performance Llama3 implementation using Micronaut and GraalVM Native Image |
|
Experimental |
| 189 |
aratan/ApiCloudLLaMA
The idea is to make an api that everyone can consume in their GPT4-like... |
|
Experimental |
| 190 |
SanMog/Uroboros
Automated red-teaming framework for LLMs. Tests GPT-4o, Claude, Llama... |
|
Experimental |
| 191 |
botosadam/matryoshka
π Build Ruby gems that utilize Rust for enhanced performance through two... |
|
Experimental |
| 192 |
Atsusheeesh/vllm-daily
π Summarize merged PRs daily with vLLM, ensuring you stay updated on key... |
|
Experimental |
| 193 |
miga1999/AirClaw
Run OpenClaw locally on any GPU or CPU without API costs, supporting large... |
|
Experimental |
| 194 |
codewithosama03/openrouter-free-model
π Explore and manage free models on OpenRouter with this web app, featuring... |
|
Experimental |
| 195 |
nherx/free-llm-api-resources
π€ Discover free API access and credits for various legitimate large language... |
|
Experimental |
| 196 |
xxxbf0222/LlamaDeck
A command-line tool for quickly managing and experimenting with multiple... |
|
Experimental |
| 197 |
llamajs/llama
A dynamic logger for the dynamic developer |
|
Experimental |
| 198 |
tbogdala/woolyrust
A high-level Rust wrapper around llama.cpp for text generation AI with LLMs. |
|
Experimental |
| 199 |
unaidedelf8777/faster-outlines
A Lazy, high throughput and blazing fast structured text generation backend. |
|
Experimental |
| 200 |
CameLLM/CameLLM
Run your favourite LLMs locally on macOS from Swift |
|
Experimental |
| 201 |
tbogdala/woolycore
The core wrapper around llama.cpp in C to provide an easy surface to build... |
|
Experimental |
| 202 |
MaoJianwei/llama.cpp-arm-armv7l-Raspberry-Pi-Release-Prebuild
On the Releases page, you can download pre-built binaries for arm, armv7l... |
|
Experimental |
| 203 |
yasir13001/MoonAI_API
This MoonAI API service built with FastAPI that calculates and provides... |
|
Experimental |
| 204 |
TimeSurgeLabs/promptproxy
Call many AIs from a single API. |
|
Experimental |
| 205 |
themaximalist/ModelDeployer
API Proxy for AI models, rate limiting, management and more! |
|
Experimental |
| 206 |
JinHanLei/LLM-Stream-Service
Streaming API and Web page for Large Language Models (Llama3) based on... |
|
Experimental |
| 207 |
iakashpaul/Ghudsavar
Ghudsavar (Horse rider) - Is a quick llama.cpp server for CPU only runtimes |
|
Experimental |
| 208 |
mkashirin/splinter
Splinter (Sequence Processing Language Interpreter) is a tree-walking... |
|
Experimental |
| 209 |
Root1V/llm-security
JWT-based authentication and authorization gateway for locally deployed LLM... |
|
Experimental |
| 210 |
gyanaranjans/llma-rust
A simple webapp to showcase the ability to write a simple chatbot webapp... |
|
Experimental |
| 211 |
kashan-alam/ai-backend-fastapi
AI-powered backend API built with FastAPI, JWT authentication, rate... |
|
Experimental |
| 212 |
Jshulgach/NeuroBridge
NeuroBridge: Where AI perception meets real-time robotics control |
|
Experimental |
| 213 |
lufixSch/auto_llama
Supercharge your local LLM |
|
Experimental |
| 214 |
antononcube/Raku-WWW-LLaMA
Raku package that provides access to the algorithms/models of (the... |
|
Experimental |
| 215 |
m9m9ra/llama.swiftui
It`s my playground to test mokpell llama swift lib |
|
Experimental |
| 216 |
numq/text-generation
JVM library for text generation, written in Kotlin and based on the C++... |
|
Experimental |
| 217 |
lenML/llama2-tokenizer.js
llama2 tokenizer for javascript |
|
Experimental |
| 218 |
tbogdala/ai_notepad
A lightweight Rust application to test interaction with large language... |
|
Experimental |
| 219 |
JavaLLM/llama4j
An easy-to-use Java SDK for running LLaMA models on edge devices, powered by... |
|
Experimental |
| 220 |
yachty66/aicomputer
Open source DIY AI computing platform: Build a powerful RTX 3090 GPU rig... |
|
Experimental |
| 221 |
coderonion/awesome-mojo-max-mlir
A collection of some awesome public MAX platform, Mojo programming language... |
|
Experimental |
| 222 |
pantaleone-ai/private-ai-stack
Deploy a complete, self-hosted AI stack for private LLMs, agentic workflows,... |
|
Experimental |
| 223 |
zTgx/llama.rust
LLM inference in Rust |
|
Experimental |
| 224 |
niansa/libjustlm
Super easy to use library for doing LLaMA/GPT-J stuff! - Mirror of:... |
|
Experimental |
| 225 |
eccenca/llama-index-cmem
llama-index tools eccenca Corporate Memory Integration |
|
Experimental |
| 226 |
georon/llama_test_proj
Skeleton project to run and test Llama and Chromadb locally on a gaming... |
|
Experimental |
| 227 |
diogok/llamautils
Some python utilities for running llama.cpp on linux |
|
Experimental |
| 228 |
Abdullahali77/AI_Testing_CLI
A specialized command-line tool that generates Python unit tests for your... |
|
Experimental |
| 229 |
tripolskypetr/agent-tune
A React-based tool for constructing fine-tuning datasets with list and grid... |
|
Experimental |
| 230 |
NavodPeiris/node_llama
run llama models using llamafile and communicate with llama models through... |
|
Experimental |
| 231 |
scttfrdmn/genkit-aws
AWS plugins for Google's GenKit framework - add AWS Bedrock models and... |
|
Experimental |
| 232 |
OnlyF0uR/interactive-ai
Rust CLI application for interacting with LLMs for Llama & OpenRouter. |
|
Experimental |
| 233 |
jazibjohar/ai-text-structor
A powerful asynchronous framework for orchestrating Large Language Model... |
|
Experimental |
| 234 |
pAI-OS/fetch_llama_cpp
llama.cpp downloader that selects the latest and best available binaries for... |
|
Experimental |
| 235 |
fasuizu-br/brainiall-llm-gateway
Brainiall LLM Gateway β 113+ AI models via OpenAI-compatible API. Claude,... |
|
Experimental |
| 236 |
aruntemme/llamacpp-swap-boilerplate
A cross-platform template for running and managing llama-swap with... |
|
Experimental |
| 237 |
NeuralWeights/Llama-Server-AuthKeys
Authorization tokens to access llama.cpp server (LM Studio, Ollama, Msty,... |
|
Experimental |
| 238 |
updcon/libmisc-clj
DKD miscellaneous for Clojure development |
|
Experimental |
| 239 |
d1pankarmedhi/Phi3-rust
Serve Phi3 with Candle and Actix π¦ |
|
Experimental |
| 240 |
Inferra/Inferra-Python-SDK
Official Python SDK for Inferra API access |
|
Experimental |
| 241 |
asaddi/lv-serve
Llama 3.2 Vision OpenAI-like API server |
|
Experimental |
| 242 |
Inferra/Inferra-JS-SDK
Official JavaScript/TypeScript SDK for Inferra API access |
|
Experimental |
| 243 |
0xricksanchez/AIonic
AIonic: A unified, user-friendly Rust library for seamless integration with... |
|
Experimental |
| 244 |
3axislabs/llm4j
Build Context Aware LLM Apps using Java |
|
Experimental |
| 245 |
shakfu/llamalib
Thin cython, pybind11, and nanobind wrappers around llama.cpp |
|
Experimental |