knowledge-to-action-mcp and mcp-rag-server
About knowledge-to-action-mcp
tac0de/knowledge-to-action-mcp
MCP server for Obsidian GraphRAG, agent-ready context, preview-only planning, and safe repo handoffs
Combines graph-aware note retrieval with optional embeddings-based semantic reranking to surface contextual neighbors, then structures results into agent-ready packets containing briefs, risks, and repo file hints. Implements preview-only action planning and bounded workspace inspection (ripgrep, git status) without exposing shell access, integrating via stdio with Claude, VS Code, and Cursor.
About mcp-rag-server
kwanLeeFrmVi/mcp-rag-server
mcp-rag-server is a Model Context Protocol (MCP) server that enables Retrieval Augmented Generation (RAG) capabilities. It empowers Large Language Models (LLMs) to answer questions based on your document content by indexing and retrieving relevant information efficiently.
Supports multiple embedding providers (OpenAI, Ollama, Granite, Nomic) with a SQLite-backed vector store, exposing indexing and retrieval operations as MCP tools and resources over stdio. Processes documents in five formats (.txt, .md, .json, .jsonl, .csv) with configurable chunking, enabling seamless integration into any MCP-compatible client or LLM application.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work