mcp-memory-service and memorix
These are **complements**: mcp-memory-service provides the persistent memory backend and knowledge graph infrastructure, while memorix implements that memory layer specifically for AI coding agents across multiple IDEs, allowing them to work together in an integrated agentic system.
About mcp-memory-service
doobidoo/mcp-memory-service
Open-source persistent memory for AI agent pipelines (LangGraph, CrewAI, AutoGen) and Claude. REST API + knowledge graph + autonomous consolidation.
Consolidates multi-agent memory using a knowledge graph with typed edges (causes, fixes, contradicts) and autonomous compression, accessible via REST API with ONNX-based embeddings that run locally. Implements Remote MCP support for browser-based claude.ai integration via Server-Sent Events, alongside traditional desktop MCP, with OAuth 2.0 authentication and self-hosted infrastructure (no cloud lock-in). Agent identity is tracked via `X-Agent-ID` headers for scoped retrieval, and conversation threading is preserved through `conversation_id` fields, enabling both shared memory across agent fleets and inter-agent messaging through semantic tag-based filtering.
About memorix
AVIDS2/memorix
Cross-Agent Memory Bridge Persistent memory for AI coding agents across 10 IDEs (Cursor, Windsurf, Claude Code, Codex, Copilot, Kiro, Antigravity, OpenCode, Trae, Gemini CLI) via MCP. Team collaboration, auto-cleanup, mini-skills, workspace sync. Never re-explain your project again.
Implements a git-aware memory pipeline that separates commit provenance from reasoning memory, storing both through an MCP server available in stdio mode (per-IDE) or HTTP mode (shared background process). Agents query memory through `memorix_search`, `memorix_timeline`, and `memorix_resolve` tools that apply source-aware retrieval and automatic compaction based on memory formation rules and project binding configuration.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work