memorix and memora
These are **complements**: AVIDS2/memorix provides IDE-agnostic persistence across coding agents while memora provides the semantic storage and knowledge graph backend that could power cross-agent memory systems like memorix.
About memorix
AVIDS2/memorix
Cross-Agent Memory Bridge Persistent memory for AI coding agents across 10 IDEs (Cursor, Windsurf, Claude Code, Codex, Copilot, Kiro, Antigravity, OpenCode, Trae, Gemini CLI) via MCP. Team collaboration, auto-cleanup, mini-skills, workspace sync. Never re-explain your project again.
Implements a git-aware memory pipeline that separates commit provenance from reasoning memory, storing both through an MCP server available in stdio mode (per-IDE) or HTTP mode (shared background process). Agents query memory through `memorix_search`, `memorix_timeline`, and `memorix_resolve` tools that apply source-aware retrieval and automatic compaction based on memory formation rules and project binding configuration.
About memora
agentic-box/memora
Give your AI agents persistent memory — MCP server for semantic storage, knowledge graphs, and cross-session context
Implements a Model Context Protocol (MCP) server with pluggable embedding backends (OpenAI, sentence-transformers, TF-IDF) and multi-tiered storage—local SQLite, Cloudflare D1, or S3/R2 with optional encryption and compression. Features include interactive knowledge graph visualization, RAG-powered chat with streaming LLM tool calling, event notifications for inter-agent communication, and automated memory deduplication via LLM comparison. Integrates with Claude Code and Codex CLI through stdio or HTTP transports.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work