memora and Memory-Plus
About memora
agentic-box/memora
Give your AI agents persistent memory โ MCP server for semantic storage, knowledge graphs, and cross-session context
Implements a Model Context Protocol (MCP) server with pluggable embedding backends (OpenAI, sentence-transformers, TF-IDF) and multi-tiered storageโlocal SQLite, Cloudflare D1, or S3/R2 with optional encryption and compression. Features include interactive knowledge graph visualization, RAG-powered chat with streaming LLM tool calling, event notifications for inter-agent communication, and automated memory deduplication via LLM comparison. Integrates with Claude Code and Codex CLI through stdio or HTTP transports.
About Memory-Plus
Yuchen20/Memory-Plus
๐ง ๐ด๐๐๐๐๐-๐ท๐๐๐ is a lightweight, local RAG memory store for MCP agents. Easily record, retrieve, update, delete, and visualize persistent "memories" across sessionsโperfect for developers working with multiple AI coders (like Windsurf, Cursor, or Copilot) or anyone who wants their AI to actually remember them.
Built on Google's Embedding API for semantic search, Memory-Plus stores encoded memories locally and supports versioning to track changes over time. It integrates as an MCP server via stdio transport, compatible with VS Code, Cursor, Cline, and other MCP-enabled IDEs, with optional resource-based prompting to control when agents access past context.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work