Ar5en1c/context-guardian
MCP server + proxy that cuts cloud LLM costs 36-42% by indexing context locally with Ollama
Stars
4
Forks
—
Language
TypeScript
License
MIT
Last pushed
Apr 09, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/Ar5en1c/context-guardian"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
SamurAIGPT/llm-wiki-agent
A personal knowledge base that builds and maintains itself. Drop in sources — Claude (or...
EverMind-AI/EverOS
A memory OS that makes your OpenClaw agents more personal while saving tokens.
alvinreal/awesome-opensource-ai
Curated list of the best truly open-source AI projects, models, tools, and infrastructure.
kongusen/loom-agent
Python agent framework with AgentHarness: planner → generator ⇄ evaluator loops for long-running...
shakfu/cyllama
A thin cython wrapper around llama.cpp, whisper.cpp and stable-diffusion.cpp