mem0 and memwire
These tools are competitors, with one offering a universal memory layer likely for various AI agents while the other focuses on a self-hosted memory infrastructure layer, suggesting differing deployment and control preferences for managing AI memory.
About mem0
mem0ai/mem0
Universal memory layer for AI Agents
Implements multi-level memory (user, session, agent state) with adaptive retrieval that achieves 26% higher accuracy and 90% lower token usage than baseline approaches. Supports multiple LLMs and vector stores, with SDKs for Python and JavaScript, plus integrations for LangGraph and CrewAI. Offers both self-hosted open-source deployment and a managed platform with CLI tooling for memory management operations.
About memwire
memoryoss/memwire
Open source self-hosted AI memory infrastructure layer
Implements graph-based semantic memory with categorized facts (preferences, events, entities, instructions) that strengthen or decay based on feedback loops, enabling persistent context recall across conversations. Provides both Python SDK and FastAPI REST interface, integrating with any LLM provider (OpenAI, Anthropic, Ollama) and vector stores (Qdrant, Pinecone, ChromaDB), while supporting multi-tenant isolation and knowledge base ingestion alongside conversation memory.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work