adrianliechti/wingman
Inference Hub for AI at Scale
Supports multi-provider LLM integration (OpenAI, Anthropic, Gemini, Bedrock, local Ollama) with pluggable document processing pipelines (extractors, segmenters, retrievers) for RAG workflows. Offers modular architecture with built-in tools, Model Context Protocol (MCP) support for external tool servers, and load balancing/rate limiting across providers. Exposes OpenAI-compatible APIs with full OpenTelemetry observability and YAML-based configuration for chains, agents, and complex AI workflows.
Stars
73
Forks
12
Language
Go
License
MIT
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/adrianliechti/wingman"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
langbot-app/LangBot
Production-grade platform for building agentic IM bots - 生产级多平台智能机器人开发平台. 提供 Agent、知识库编排、插件系统 /...
open-webui/open-webui
User-friendly AI Interface (Supports Ollama, OpenAI API, ...)
cactus-compute/cactus
Low-latency AI engine for mobile devices & wearables
rudrankriyam/Foundation-Models-Framework-Example
Example apps for Foundation Models Framework in iOS 26 and macOS 26
sigoden/aichat
All-in-one LLM CLI tool featuring Shell Assistant, Chat-REPL, RAG, AI Tools & Agents, with...