nocturne_memory and memorix
These are complements: nocturne_memory provides the persistent memory backend/server infrastructure while memorix acts as a cross-IDE client bridge to connect multiple AI agents to shared memory, requiring a memory system like nocturne_memory to store and retrieve collaborative state.
About nocturne_memory
Dataojitori/nocturne_memory
A lightweight, rollbackable, and visual Long-Term Memory Server for MCP Agents. Say goodbye to Vector RAG and amnesia. Empower your AI with persistent, graph-like structured memory across any model, session, or tool. Drop-in replacement for OpenClaw.
Implements a graph-based memory architecture with SQLite/PostgreSQL backends, where AI agents can create, update, and rollback their own structured memories through MCP—eliminating vector RAG's semantic lossy compression and enabling condition-triggered disclosure of hierarchical knowledge graphs with human-auditable versioning. Includes a visual dashboard for memory exploration, diff review, and governance; integrates natively with Claude Desktop, Cursor, and other MCP-compatible frameworks as a direct OpenClaw replacement.
About memorix
AVIDS2/memorix
Cross-Agent Memory Bridge Persistent memory for AI coding agents across 10 IDEs (Cursor, Windsurf, Claude Code, Codex, Copilot, Kiro, Antigravity, OpenCode, Trae, Gemini CLI) via MCP. Team collaboration, auto-cleanup, mini-skills, workspace sync. Never re-explain your project again.
Implements a git-aware memory pipeline that separates commit provenance from reasoning memory, storing both through an MCP server available in stdio mode (per-IDE) or HTTP mode (shared background process). Agents query memory through `memorix_search`, `memorix_timeline`, and `memorix_resolve` tools that apply source-aware retrieval and automatic compaction based on memory formation rules and project binding configuration.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work