memorix and MegaMemory

These tools are ecosystem siblings: one is a cross-agent memory bridge providing persistent memory across multiple IDEs, while the other is a persistent project knowledge graph server that could serve as a backend for such a bridge, offering semantic search and in-process embeddings.

memorix
56
Established
MegaMemory
50
Established
Maintenance 13/25
Adoption 10/25
Maturity 20/25
Community 13/25
Maintenance 10/25
Adoption 8/25
Maturity 20/25
Community 12/25
Stars: 208
Forks: 18
Downloads:
Commits (30d): 0
Language: TypeScript
License: Apache-2.0
Stars: 59
Forks: 7
Downloads:
Commits (30d): 0
Language: TypeScript
License: MIT
No risk flags
No risk flags

About memorix

AVIDS2/memorix

Cross-Agent Memory Bridge Persistent memory for AI coding agents across 10 IDEs (Cursor, Windsurf, Claude Code, Codex, Copilot, Kiro, Antigravity, OpenCode, Trae, Gemini CLI) via MCP. Team collaboration, auto-cleanup, mini-skills, workspace sync. Never re-explain your project again.

Implements a git-aware memory pipeline that separates commit provenance from reasoning memory, storing both through an MCP server available in stdio mode (per-IDE) or HTTP mode (shared background process). Agents query memory through `memorix_search`, `memorix_timeline`, and `memorix_resolve` tools that apply source-aware retrieval and automatic compaction based on memory formation rules and project binding configuration.

About MegaMemory

0xK3vin/MegaMemory

Persistent project knowledge graph for coding agents. MCP server with semantic search, in-process embeddings, and web explorer.

Uses in-process ONNX embeddings (all-MiniLM-L6-v2) and SQLite with WAL for zero-dependency semantic search and persistence. Operates as an MCP stdio server integrated with Claude Code, OpenCode, Antigravity, and Codex, with built-in two-way merge conflict resolution for collaborative knowledge graph management across branches. The LLM itself acts as the indexer—concepts are stored in natural language rather than parsed code symbols—enabling agents to update the graph after each task and query semantic context before starting new ones.

Scores updated daily from GitHub, PyPI, and npm data. How scores work