MemPalace
A local, open-source AI memory system that stores all your conversations verbatim in ChromaDB and makes them findable via semantic search, achieving 96.6% LongMemEval R@5 with zero API calls.
At a Glance
Fully free and open-source under the MIT License. No subscription, no cloud, no API key required.
Engagement
Available On
Alternatives
Listed Apr 2026
About MemPalace
MemPalace is a free, open-source AI memory system that stores every conversation, decision, and debugging session verbatim in a local ChromaDB instance — nothing is summarized or discarded. It achieves the highest published LongMemEval R@5 score (96.6% in raw mode, 100% with optional reranking) without requiring any API key or cloud service. The system organizes memories into a hierarchical "palace" structure — wings, halls, rooms, closets, and drawers — inspired by the ancient method of loci, giving AI agents a navigable map rather than a flat search index. It integrates with Claude Code, Claude, ChatGPT, Cursor, Gemini, and local models via an MCP server exposing 19 tools.
- Palace structure — Memories are organized into wings (people/projects), halls (memory types), and rooms (specific topics), delivering a +34% retrieval improvement over unfiltered search.
- Raw verbatim storage — All content is stored as-is in ChromaDB without LLM summarization, which is the source of the 96.6% benchmark score; run
pip install mempalaceandmempalace init <dir>to get started. - MCP server with 19 tools — Connect once via
claude mcp add mempalace -- python -m mempalace.mcp_serverand your AI automatically callsmempalace_search,mempalace_kg_query, and other tools on demand. - Temporal knowledge graph — SQLite-backed entity-relationship triples with validity windows let you query what was true at any point in time, similar to Zep's Graphiti but fully local and free.
- AAAK compression dialect (experimental) — A lossy abbreviation system for packing repeated entities into fewer tokens at scale; currently regresses LongMemEval vs raw mode (84.2% vs 96.6%) and is not the storage default.
- Specialist agents — Define focused agents (reviewer, architect, ops) each with their own wing and AAAK diary in the palace; agents persist expertise across sessions without bloating your config.
- Auto-save hooks — Claude Code hooks trigger structured saves every 15 messages and before context compaction, keeping the memory palace current automatically.
- Three mining modes —
projectsfor code and docs,convosfor chat exports (Claude, ChatGPT, Slack), andgeneralfor auto-classification into decisions, preferences, milestones, problems, and emotional context. - Fully local — Requires only Python 3.9+, ChromaDB, and PyYAML; no internet connection after install, no data leaves your machine.
Community Discussions
Be the first to start a conversation about MemPalace
Share your experience with MemPalace, ask questions, or help others learn from your insights.
Pricing
Open Source (MIT)
Fully free and open-source under the MIT License. No subscription, no cloud, no API key required.
- Verbatim conversation storage in ChromaDB
- Palace structure (wings, halls, rooms, closets, drawers)
- MCP server with 19 tools
- Temporal knowledge graph (SQLite)
- AAAK compression dialect
Capabilities
Key Features
- Verbatim conversation storage in ChromaDB
- Palace structure: wings, halls, rooms, closets, drawers
- 96.6% LongMemEval R@5 in raw mode (zero API calls)
- 100% LongMemEval R@5 with optional Haiku reranking
- MCP server with 19 tools for Claude, ChatGPT, Cursor, Gemini
- Temporal knowledge graph (SQLite) with validity windows
- AAAK lossy compression dialect for token-efficient context loading
- 4-layer memory stack (L0 identity, L1 critical facts, L2 room recall, L3 deep search)
- Specialist agents with per-agent wings and AAAK diaries
- Auto-save hooks for Claude Code (save + precompact)
- Three mining modes: projects, convos, general
- Semantic search with wing/room metadata filtering
- Mega-file splitter for concatenated transcripts
- Contradiction detection utility (fact_checker.py, experimental)
- Fully local — no API key, no cloud, no data egress
- Gemini CLI native integration
- Wake-up command for local model context loading
- Python API for programmatic access
