Back to all articles
15 MIN READ

MemPalace: The Open-Source AI Memory System That Scores 96.6% — Complete Guide (2026)

By Learnia Team

MemPalace: The Open-Source AI Memory System That Scores 96.6%

This article is written in English. Our training modules are available in multiple languages.

📅 Last Updated: April 8, 2026 — MemPalace v3.0.0, released April 5, 2026.

📚 Related: Claude Code Complete Guide | MCP Protocol Guide | Best AI Coding Tools 2026

Every conversation you have with an AI disappears when the session ends. Six months of decisions, debugging sessions, architecture debates — gone. You start over every time.

Other memory systems try to fix this by letting AI decide what's worth remembering. The AI extracts "user prefers Postgres" and throws away the conversation where you explained why. You lose the reasoning, the context, the nuance.

MemPalace takes a different approach: store everything, then make it findable. On April 5, 2026, actress and tech entrepreneur Milla Jovovich and developer Ben Sigman released MemPalace v3.0.0 — an open-source AI memory system that has already hit 21.7K GitHub stars in three days. Here's why the developer community is paying attention.


The Problem MemPalace Solves

After six months of daily AI use, you've generated approximately 19.5 million tokens of conversations. That's every decision, every solution, every trade-off analysis. The question is: where did it all go?

ApproachToken CostAnnual CostRecall
Paste everything into context19.5M — doesn't fitImpossibleN/A
LLM summarization~650K tokens~$507/yearLossy
MemPalace wake-up~170 tokens~$0.70/year96.6%
MemPalace + 5 searches~13,500 tokens~$10/year96.6%

The difference is dramatic: $10/year for 96.6% recall vs. $507/year for lossy summaries that throw away context.


The Palace Architecture

MemPalace is inspired by the ancient Greek Method of Loci — a memorization technique where you place ideas in rooms of an imaginary building and walk through it to recall them. MemPalace applies the same principle to AI memory:

Loading diagram…

The Building Blocks

ElementWhat It IsExample
WingA person or projectwing_kai, wing_driftwood
RoomA specific topic inside a wingauth-migration, graphql-switch
HallA memory type connecting roomshall_facts, hall_events, hall_discoveries
TunnelCross-wing connections (same topic, different contexts)Auth-migration appears in both Person and Project wings
ClosetA summary that points to original filesPlain-text summaries (AAAK coming soon)
DrawerThe original verbatim fileYour exact conversation, never summarized

Why Structure Matters

This isn't cosmetic organization. The palace structure directly improves retrieval accuracy:

Search modeRecall (R@10)Improvement
Search all closets (unfiltered)60.9%Baseline
Search within wing73.1%+12%
Search wing + hall84.8%+24%
Search wing + room94.8%+34%

The architecture gives AI a navigable map instead of a flat search index. When you ask "what did we decide about auth?", MemPalace knows which wing to search, which rooms to check, and which halls to follow, instead of scanning every piece of data.


The Memory Stack

MemPalace uses a 4-layer system that loads only what's needed:

LayerContentSizeWhen Loaded
L0Identity — who is this AI?~50 tokensAlways
L1Critical facts — team, projects, preferences~120 tokens (AAAK)Always
L2Room recall — recent sessions, current projectOn demandWhen topic comes up
L3Deep search — semantic query across all closetsOn demandWhen explicitly asked

Your AI wakes up with L0 + L1 (~170 tokens) and immediately knows your world — your team members, your projects, your preferences. Searches only fire when needed, keeping costs near zero.


Getting Started

Installation

pip install mempalace

Requirements: Python 3.9+, no API key, no internet after install.

Setup

# Initialize — guided onboarding, sets up wings for your people and projects
mempalace init ~/projects/myapp

# Mine your project files (code, docs, notes)
mempalace mine ~/projects/myapp

# Mine your AI conversations (Claude, ChatGPT, Slack exports)
mempalace mine ~/chats/ --mode convos

# General mode — auto-classifies into decisions, milestones, problems
mempalace mine ~/chats/ --mode convos --extract general
# Search everything
mempalace search "why did we switch to GraphQL"

# Search within a specific project
mempalace search "database decision" --wing myapp

# Search a specific topic
mempalace search "auth approach" --room auth-migration

Connect to Your AI

For Claude, ChatGPT, Cursor, Gemini (MCP-compatible tools):

claude mcp add mempalace -- python -m mempalace.mcp_server

Now your AI has 19 tools available through MCP. Ask it anything:

"What did we decide about auth last month?"

Claude calls mempalace_search automatically, gets verbatim results, and answers. You never type a search command — the AI handles it.

For local models (Llama, Mistral):

mempalace wake-up > context.txt
# Paste context.txt into your model's system prompt

MemPalace vs. The Competition

The Fundamental Difference

Most memory systems use an AI to extract what it thinks is important from your conversations — "user prefers Postgres," "team decided to use Clerk." Then they throw away the original conversation.

MemPalace stores everything verbatim and uses semantic search to find what's relevant. The 96.6% recall proves this works better than extraction-based approaches. You never lose the reasoning, the context, or the nuance behind a decision.


The 19 MCP Tools

When connected via MCP, your AI gets access to a comprehensive toolkit:

Palace (Read)

ToolPurpose
mempalace_statusPalace overview + AAAK spec
mempalace_list_wingsAll wings with counts
mempalace_list_roomsRooms within a wing
mempalace_get_taxonomyFull wing → room → count tree
mempalace_searchSemantic search with filters
mempalace_check_duplicateCheck before filing
mempalace_get_aaak_specAAAK dialect reference

Palace (Write)

ToolPurpose
mempalace_add_drawerFile new verbatim content
mempalace_delete_drawerRemove by ID

Knowledge Graph

ToolPurpose
mempalace_kg_queryEntity relationships with time
mempalace_kg_addAdd facts
mempalace_kg_invalidateEnd facts (temporal validity)
mempalace_kg_timelineChronological entity story
mempalace_kg_statsGraph overview
ToolPurpose
mempalace_traverseWalk the graph across wings
mempalace_find_tunnelsRooms bridging two wings
mempalace_graph_statsConnectivity overview
mempalace_diary_writeAgent diary entry (AAAK)
mempalace_diary_readRead recent diary entries

Specialist Agents

MemPalace lets you create agents with their own memory:

~/.mempalace/agents/
  ├── reviewer.json       # code quality, patterns, bugs
  ├── architect.json      # design decisions, tradeoffs
  └── ops.json            # deploys, incidents, infra

Each agent has a focus, keeps a diary (in AAAK), and builds expertise by reading its own history. The reviewer remembers every bug pattern it's seen. The architect remembers every design decision. They're specialist lenses on your data.

Your CLAUDE.md only needs one line:

You have MemPalace agents. Run mempalace_list_agents to see them.

Honest Limitations

MemPalace launched with impressive numbers but also with acknowledged issues. To their credit, the creators published a detailed correction on April 7:

What Was Overstated

  • AAAK compression — Originally claimed "30×" compression. Actually lossy, and small-scale examples showed no token savings
  • "+34% palace boost" — Compares unfiltered to filtered search. Real but not a novel technique — standard ChromaDB metadata filtering
  • "100% with Haiku rerank" — Real result, but the rerank pipeline wasn't in the public benchmark scripts
  • Contradiction detection — Exists as fact_checker.py but not yet wired into knowledge graph operations

What's Real

  • 96.6% LongMemEval — Raw mode, 500 questions, independently reproduced
  • Local, free, no cloud — Actually runs entirely on your machine
  • Palace architecture — Real and useful for retrieval, even if the boost is standard filtering
  • 19 MCP tools — Functional and well-documented

Who Should Use MemPalace?

Ideal Users

  • Solo developers managing multiple projects who lose context between sessions
  • Team leads who need to recall what decisions were made and by whom
  • Heavy AI users (Claude, ChatGPT, Cursor) who generate thousands of conversations
  • Privacy-conscious organizations that can't send data to cloud memory services

Not Ideal For

  • Casual AI users who have a few conversations per week — the setup overhead isn't justified
  • Teams needing real-time collaboration — MemPalace is local to each machine
  • Windows users — Some issues reported with macOS ARM64; Windows support is less tested (early days)

Conclusion

MemPalace proves that AI memory doesn't need to be expensive, cloud-dependent, or lossy. By storing everything raw and making it navigable through the palace architecture, it achieves the highest recall scores ever published — for free.

The project is three days old. It has rough edges, overstated initial claims, and open issues. But the core approach — verbatim storage, structured navigation, local-first — is sound, the benchmark results are independently verified, and the team is fixing problems in real-time.

If you use AI daily and have ever wished it could remember what you discussed last month, MemPalace is worth 5 minutes of setup. Your future self — the one looking for "why did we choose Postgres?" six months from now — will thank you.

GitHub: github.com/milla-jovovich/mempalace


Newsletter

Weekly AI Insights

Tools, techniques & news — curated for AI practitioners. Free, no spam.

Free, no spam. Unsubscribe anytime.

FAQ

What is MemPalace?+

MemPalace is a free, open-source AI memory system that stores your conversations and project data locally using ChromaDB. It scored 96.6% on LongMemEval — the highest result ever published for an AI memory system — without requiring any API key or cloud service.

How does MemPalace work?+

MemPalace organizes your data into a palace architecture: wings (people/projects), rooms (topics), closets (summaries), and drawers (verbatim files). Your AI loads a tiny 170-token wake-up context, then searches the palace only when needed, keeping costs near zero.

Is MemPalace free?+

Yes. MemPalace is MIT-licensed, runs entirely on your machine, and requires no API key, no cloud service, and no subscription. The total cost of operation is approximately $10/year for search queries, compared to $507/year for LLM-based summarization approaches.

How does MemPalace compare to Mem0 and Zep?+

MemPalace scores 96.6% on LongMemEval vs ~85% for both Mem0 and Zep. MemPalace is free and local; Mem0 costs $19-249/month and Zep costs $25+/month. Both competitors require cloud services and API calls.

What is the AAAK dialect?+

AAAK is an experimental lossy abbreviation system designed to compress repeated entities into fewer tokens. It is readable by any LLM without a decoder. Currently, AAAK scores lower than raw mode (84.2% vs 96.6% on LongMemEval) and is being improved.

Does MemPalace work with Claude, ChatGPT, and other AI tools?+

Yes. MemPalace includes an MCP server with 19 tools that works with Claude, ChatGPT, Cursor, Gemini CLI, and any MCP-compatible tool. For local models (Llama, Mistral), it provides a wake-up context file and CLI search commands.

Who created MemPalace?+

MemPalace was created by Milla Jovovich and Ben Sigman. It launched as v3.0.0 on April 5, 2026, and quickly reached 21.7K GitHub stars. The project has 11 contributors and is MIT-licensed.