mcp-rag-server and mcp-ragchat
About mcp-rag-server
kwanLeeFrmVi/mcp-rag-server
mcp-rag-server is a Model Context Protocol (MCP) server that enables Retrieval Augmented Generation (RAG) capabilities. It empowers Large Language Models (LLMs) to answer questions based on your document content by indexing and retrieving relevant information efficiently.
Supports multiple embedding providers (OpenAI, Ollama, Granite, Nomic) with a SQLite-backed vector store, exposing indexing and retrieval operations as MCP tools and resources over stdio. Processes documents in five formats (.txt, .md, .json, .jsonl, .csv) with configurable chunking, enabling seamless integration into any MCP-compatible client or LLM application.
About mcp-ragchat
gogabrielordonez/mcp-ragchat
MCP server that adds RAG-powered AI chat to any website. One command from Claude Code. Local vector store, multi-provider LLM (OpenAI/Anthropic/Gemini). Zero cloud dependency.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work