memorix and MegaMemory
These tools are ecosystem siblings: one is a cross-agent memory bridge providing persistent memory across multiple IDEs, while the other is a persistent project knowledge graph server that could serve as a backend for such a bridge, offering semantic search and in-process embeddings.
About memorix
AVIDS2/memorix
Cross-Agent Memory Bridge Persistent memory for AI coding agents across 10 IDEs (Cursor, Windsurf, Claude Code, Codex, Copilot, Kiro, Antigravity, OpenCode, Trae, Gemini CLI) via MCP. Team collaboration, auto-cleanup, mini-skills, workspace sync. Never re-explain your project again.
Implements a git-aware memory pipeline that separates commit provenance from reasoning memory, storing both through an MCP server available in stdio mode (per-IDE) or HTTP mode (shared background process). Agents query memory through `memorix_search`, `memorix_timeline`, and `memorix_resolve` tools that apply source-aware retrieval and automatic compaction based on memory formation rules and project binding configuration.
About MegaMemory
0xK3vin/MegaMemory
Persistent project knowledge graph for coding agents. MCP server with semantic search, in-process embeddings, and web explorer.
Uses in-process ONNX embeddings (all-MiniLM-L6-v2) and SQLite with WAL for zero-dependency semantic search and persistence. Operates as an MCP stdio server integrated with Claude Code, OpenCode, Antigravity, and Codex, with built-in two-way merge conflict resolution for collaborative knowledge graph management across branches. The LLM itself acts as the indexer—concepts are stored in natural language rather than parsed code symbols—enabling agents to update the graph after each task and query semantic context before starting new ones.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work