enhanced-mcp-memory and MegaMemory
One project is a fork of the other, extending the original Model Context Protocol (MCP) server with enhanced memory and task management features, making them ecosystem siblings.
About enhanced-mcp-memory
cbunting99/enhanced-mcp-memory
An enhanced MCP (Model Context Protocol) server for intelligent memory and task management, designed for AI assistants and development workflows. Features semantic search, automatic task extraction, knowledge graphs, and comprehensive project management.
Implements a 5-stage sequential thinking engine with token optimization (30-70% compression) and automatic context summarization for conversation continuity, built on FastMCP with SQLite persistence. Automatically detects project conventions (OS, package managers, build tools, runtime types) and learns command patterns to correct AI suggestions—particularly useful for cross-platform development. Exposes 20+ MCP tools including thinking chains, task decomposition, and knowledge graph relationships that link memories to code file paths.
About MegaMemory
0xK3vin/MegaMemory
Persistent project knowledge graph for coding agents. MCP server with semantic search, in-process embeddings, and web explorer.
Uses in-process ONNX embeddings (all-MiniLM-L6-v2) and SQLite with WAL for zero-dependency semantic search and persistence. Operates as an MCP stdio server integrated with Claude Code, OpenCode, Antigravity, and Codex, with built-in two-way merge conflict resolution for collaborative knowledge graph management across branches. The LLM itself acts as the indexer—concepts are stored in natural language rather than parsed code symbols—enabling agents to update the graph after each task and query semantic context before starting new ones.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work