The complete story of how MemPalace was created, who built it, how it works, and why it scored higher than every paid competitor.
Best known as Alice in the Resident Evil franchise and Leeloo in The Fifth Element, Milla Jovovich had been using AI tools daily for months before starting MemPalace. She accumulated thousands of conversations with ChatGPT and Claude — decisions, creative brainstorms, business strategy, debugging sessions — and noticed the same pattern every time: every session began from zero.
She tried every existing memory tool and found the same fundamental flaw in all of them: they used an LLM to decide what was worth remembering, and consistently discarded the reasoning she needed most. Her GitHub is at github.com/milla-jovovich.
Developer and co-author. Ben Sigman (@bensig) brought the engineering architecture to MemPalace — the ChromaDB storage layer, the knowledge graph SQLite design, the AAAK compression dialect, and the 19-tool MCP server.
His launch tweet — "My friend Milla Jovovich and I spent months building MemPalace with Claude Code. First perfect score on LongMemEval. 5,400 GitHub stars in 24 hours." — accumulated over 1.5 million impressions and triggered a wave of community analysis that helped improve the project.
The core frustration was this: six months of daily AI use produces approximately 19.5 million tokens of conversations — every architecture decision, every debugging session, every "we tried X and it failed because Y" — all trapped in chat windows that disappear when the session ends.
Existing tools like Mem0 and Zep tried to solve this by having an LLM extract "key facts" and discard the rest. This sounds efficient but has a fatal flaw: the reasoning — the why behind the decision — is exactly what gets discarded. You're left with "user prefers Postgres" but not the conversation where you explained that it was because your dataset would exceed 10GB and you needed concurrent writes.
"I wanted my AI to remember the way I remember — not just the conclusions, but the journey. The alternatives I considered, the reasons I changed my mind, the nuance. That's what existing memory systems throw away."
— Milla Jovovich, on the motivation behind MemPalace
The fundamental architectural decision in MemPalace is the opposite of every other tool: don't summarise, don't extract, don't decide what matters — store every word verbatim in ChromaDB and use semantic vector search to find it.
The name and structure came from the ancient Greek technique of memorising speeches by mentally placing ideas in rooms of an imaginary building — the "method of loci" or "memory palace." Walk through the building, find the idea. MemPalace applies this exact structure to AI memory:
wing_kai, wing_driftwood)auth-migration, graphql-switch)MemPalace was built over several months using Claude Code — Anthropic's AI-assisted coding tool. Milla Jovovich described the collaboration: her vision for how AI memory should work, combined with Ben Sigman's engineering expertise to design the architecture, implement ChromaDB storage, write the MCP server, and build the AAAK compression dialect.
The repository (github.com/milla-jovovich/mempalace) is written in Python (98.6%) with Shell scripts (1.4%) and contains:
| File / Module | Purpose |
|---|---|
cli.py | CLI entry point — all mempalace commands |
mcp_server.py | MCP server with 19 tools, AAAK auto-teach, memory protocol |
knowledge_graph.py | Temporal entity-relationship graph (SQLite) |
palace_graph.py | Room-based navigation graph |
dialect.py | AAAK compression dialect |
miner.py | Project file ingest |
convo_miner.py | Conversation ingest — chunks by exchange pair |
searcher.py | Semantic search via ChromaDB |
onboarding.py | Guided setup — generates AAAK bootstrap + wing config |
layers.py | 4-layer memory stack (L0–L3) |
hooks/mempal_save_hook.sh | Auto-save every 15 messages in Claude Code |
hooks/mempal_precompact_hook.sh | Emergency save before context compression |
benchmarks/longmemeval_bench.py | Reproducible LongMemEval runner |
MemPalace v3.0.0 launched on April 6, 2026. Within 24 hours it had accumulated 5,400 GitHub stars; by 72 hours, 26,900. Ben Sigman's launch tweet hit 1.5 million+ impressions. The story caught fire: an A-list actress with a GitHub account had co-built the highest-scoring free AI memory tool ever benchmarked. HackerNews's top comment: "Missed opportunity to call it Resident Eval."
The community also caught real problems in the original README within hours — overstated AAAK compression claims, a misleading token count example, an ambiguous "+34% palace boost" stat. The team responded the same day with a transparent "Note from Milla & Ben — April 7, 2026" that addressed every criticism directly:
len(text)//3) instead of a real tokenizer — correctedThat honesty — rare in the world of AI launch announcements — turned critics into contributors. The project now has 14 contributors and 125+ commits.
Key contributors include @bensig (Ben Sigman, co-author), @milla-jovovich (Milla Jovovich, co-author), @igorls, @adv3nt3, @sheetsync, and @ac-opensource. Community members who caught critical bugs include @panuhorsmalahti, @lhl, and @gizmax (who independently reproduced the 96.6% benchmark on an M2 Ultra in under 5 minutes).
View All Contributors ↗