Lyellr88/MARM-Systems

Turn AI into a persistent, memory-powered collaborator. Universal MCP Server (supports HTTP, STDIO, and WebSocket) enabling cross-platform AI memory, multi-agent coordination, and context sharing. Built with MARM protocol for structured reasoning that evolves with your work.

54
/ 100
Established

# Technical Summary Implements semantic vector-based memory indexing with auto-classification of conversation content (code, decisions, configs) and enables cross-session recall via FastAPI-backed HTTP/STDIO transports that integrate natively with Claude, Gemini, and other MCP-compatible agents. The architecture uses SQLite with WAL mode for persistent storage and connection pooling, exposing 18 MCP tools for granular memory control—including structured session logs, reusable notebooks, and smart context fallbacks when vector similarity alone is insufficient. Designed for production workflows requiring reliable long-term context across multiple AI agents and deployment cycles, with Docker containerization and rate-limiting built-in.

251 stars.

No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 15 / 25
Community 19 / 25

How are scores calculated?

Stars

251

Forks

42

Language

Python

License

MIT

Last pushed

Jan 24, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mcp/Lyellr88/MARM-Systems"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.