mcp-memory-service and MARM-Systems

These are competitors offering similar core functionality—both provide persistent memory systems for multi-agent AI frameworks via MCP servers—though one emphasizes knowledge graph consolidation while the other prioritizes transport protocol flexibility and cross-platform coordination.

mcp-memory-service
73
Verified
MARM-Systems
48
Emerging
Maintenance 25/25
Adoption 10/25
Maturity 16/25
Community 22/25
Maintenance 10/25
Adoption 10/25
Maturity 9/25
Community 19/25
Stars: 1,504
Forks: 215
Downloads:
Commits (30d): 153
Language: Python
License: Apache-2.0
Stars: 251
Forks: 42
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No Package No Dependents
No Package No Dependents

About mcp-memory-service

doobidoo/mcp-memory-service

Open-source persistent memory for AI agent pipelines (LangGraph, CrewAI, AutoGen) and Claude. REST API + knowledge graph + autonomous consolidation.

Consolidates multi-agent memory using a knowledge graph with typed edges (causes, fixes, contradicts) and autonomous compression, accessible via REST API with ONNX-based embeddings that run locally. Implements Remote MCP support for browser-based claude.ai integration via Server-Sent Events, alongside traditional desktop MCP, with OAuth 2.0 authentication and self-hosted infrastructure (no cloud lock-in). Agent identity is tracked via `X-Agent-ID` headers for scoped retrieval, and conversation threading is preserved through `conversation_id` fields, enabling both shared memory across agent fleets and inter-agent messaging through semantic tag-based filtering.

About MARM-Systems

Lyellr88/MARM-Systems

Turn AI into a persistent, memory-powered collaborator. Universal MCP Server (supports HTTP, STDIO, and WebSocket) enabling cross-platform AI memory, multi-agent coordination, and context sharing. Built with MARM protocol for structured reasoning that evolves with your work.

# Technical Summary Implements semantic vector-based memory indexing with auto-classification of conversation content (code, decisions, configs) and enables cross-session recall via FastAPI-backed HTTP/STDIO transports that integrate natively with Claude, Gemini, and other MCP-compatible agents. The architecture uses SQLite with WAL mode for persistent storage and connection pooling, exposing 18 MCP tools for granular memory control—including structured session logs, reusable notebooks, and smart context fallbacks when vector similarity alone is insufficient. Designed for production workflows requiring reliable long-term context across multiple AI agents and deployment cycles, with Docker containerization and rate-limiting built-in.

Scores updated daily from GitHub, PyPI, and npm data. How scores work