octocode-mcp and code-memory
These are **competitors** offering different trade-offs: one prioritizes real-time search across distributed repositories with cloud-based semantic indexing, while the other emphasizes local vector indexing and offline-first codebase analysis.
About octocode-mcp
bgauryy/octocode-mcp
MCP server for semantic code research and context generation on real-time using LLM patterns | Search naturally across public & private repos based on your permissions | Transform any accessible codebase/s into AI-optimized knowledge on simple and complex flows | Find real implementations and live docs from anywhere
Implements MCP (Model Context Protocol) with LSP-powered code intelligence (Go to Definition, Find References, Call Hierarchy) across GitHub, GitLab, and local codebases, enabling compiler-level understanding without parsing. Provides modular Agent Skills—including multi-phase research with session persistence, AST-driven code analysis, dependency graphing, and PR review across seven domains—composable via CLI or direct integration into Claude/Cursor.
About code-memory
kapillamba4/code-memory
MCP server with local vector search for your codebase. Smart indexing, semantic search, Git history — all offline.
Implements a three-tier retrieval strategy (BM25 + dense vectors via SQLite-vec + AST parsing) across specialized tools that route queries by type—code definitions, architectural documentation, and Git history—powered by local sentence-transformers embeddings. Integrates as an MCP server with Claude Desktop, VS Code (Copilot/Continue), Gemini CLI, and Claude Code, supporting 8 languages with full AST extraction and 9+ with fallback whole-file indexing.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work