Back to Blog
BlogApril 7, 20264

What Is MemPalace? The AI Memory System by Milla Jovovich That Achieves 100% on LongMemEval

What Is MemPalace? The AI Memory System by Milla Jovovich That Achieves 100% on LongMemEval

Key Takeaways

  • MemPalace is a free, open-source agentic memory system that gives AI models persistent, structured, and lossless long-term memory using a virtual memory palace architecture inspired by the ancient method of loci.
  • Developed by actress Milla Jovovich and engineer Ben Sigman, it scores 100% hybrid and the highest raw results on the industry-standard LongMemEval benchmark.
  • Fully local and offline-first: runs on your machine with ChromaDB, SQLite, and filesystem—no cloud, no API costs, no data sent externally.
  • Employs hierarchical organization (wings, rooms, halls, tunnels, closets, drawers) plus AAAK lossless compression for up to 30x efficiency while keeping content fully readable.
  • Analysis of benchmarks shows it outperforms paid tools like Mem0, Zep, and others in recall accuracy and cost.
  • Community and early user feedback highlights its effectiveness for developers, teams, and local LLM users facing AI “amnesia.”

What Is MemPalace?

MemPalace is an innovative AI memory system designed to solve the persistent problem of large language models forgetting previous conversations, decisions, and project details once the context window resets. Instead of relying on lossy summarization or expensive cloud storage, it builds a structured virtual memory palace where every piece of information is stored verbatim and made instantly retrievable.

Launched in early April 2026, the project is fully open-source on GitHub under the repository milla-jovovich/mempalace. It draws direct inspiration from the classical memory palace (method of loci) technique, adapting spatial organization principles to digital knowledge management.

Benchmarks indicate MemPalace delivers the highest published scores on LongMemEval, making it a standout solution for anyone building or using long-running AI agents and applications.

Origins and Inspiration

Milla Jovovich developed the core architecture after experiencing frustration with AI tools losing critical details despite meticulous filing. Drawing from historical accounts of ancient Greek orators and modern memory champions who use spatial memory techniques to recall vast amounts of information, she envisioned a “virtual Memory Palace” for AI.

Collaborating with technical co-founder Ben Sigman, the team created a system that organizes data into familiar spatial metaphors: different wings for projects or people, rooms for topics, halls for decision types, and more. This approach leverages the brain’s natural strength in spatial navigation while providing a practical framework for AI retrieval.

The result is a system that stores everything without summarization loss, making retrieval more intuitive and accurate than keyword or flat vector search alone.

How MemPalace Works

MemPalace combines hierarchical spatial structuring, semantic search, and efficient compression in a multi-layer memory stack:

  • Mining Phase: Automatically ingests chats, code projects, Slack histories, or other data sources and classifies them into the palace structure.
  • AAAK Compression: A custom lossless format reduces token count dramatically (e.g., 1,000 tokens to approximately 120) while remaining fully readable by humans and any LLM—no special decoder needed.
  • Palace Architecture:
    • Wings: High-level categories (projects, people)
    • Rooms: Specific topics or conversations
    • Halls: Fact or event types (decisions, events)
    • Tunnels: Cross-references between elements
    • Closets/Drawers: Compressed summaries and original verbatim files
  • Storage and Retrieval: Uses ChromaDB for vector embeddings, SQLite for a temporal knowledge graph (with validity windows for facts), and local filesystem for raw data. Retrieval applies hierarchical filtering for precision.
  • Integration: Supports MCP (Model Context Protocol) for seamless use with Claude or ChatGPT, plus direct Python/CLI access for local models like Ollama.

This design ensures zero data loss and enables both semantic and structured queries.

Key Features and Technical Details

  • Lossless & Efficient: Up to 30x compression; full histories fit into tiny wake-up contexts.
  • Hierarchical Filtering: Delivers significant retrieval improvements (benchmarks show gains from wing + room filtering).
  • Temporal Knowledge Graph: Tracks entity relationships and changes over time, supporting historical and contradiction-aware queries.
  • Offline-First: One dependency; runs completely locally with no recurring costs.
  • MCP Tools: Built-in tools for agents including search, graph queries, and more.
  • Specialist Support: Domain-specific wings and auto-save capabilities for teams or complex projects.

Benchmarks and Performance

MemPalace sets a new standard according to published results:

  • LongMemEval: Highest raw score published and 100% in hybrid mode.
  • It outperforms several commercial memory systems while remaining completely free and local.

Comparisons show clear advantages in accuracy, privacy, and cost over tools requiring API calls or subscriptions. The hierarchical approach contributes measurably to precision beyond raw embeddings alone.

Getting Started with MemPalace

Installation and basic usage are straightforward via the CLI:

pip install mempalace
mempalace init ~/my-palace
mempalace mine ~/projects --mode projects
mempalace mine ~/chats --mode convos
mempalace search "decision details"
mempalace status

For Claude integration, add the MCP server. Local LLM users can generate wake-up contexts directly. The repository provides full documentation and examples.

Advanced Tips and Use Cases

  • Team Memory: Mine shared communication histories into dedicated wings for consistent organizational recall.
  • Custom Configuration: Adjust wing mappings for automatic classification of repositories or individuals.
  • Knowledge Graph Queries: Leverage temporal queries for timeline-aware information retrieval.
  • Massive Scale: Handles millions of tokens efficiently without performance degradation.
  • Hybrid Workflows: Combine with local models for extremely low ongoing costs.

Developers building long-term agents and teams managing complex projects benefit most from the structured, searchable memory.

Common Pitfalls and Solutions

  • Initial Indexing Time: Large datasets take time to mine—begin with targeted folders.
  • Retrieval Precision: Rely on hierarchical filters rather than raw search for optimal results.
  • AAAK Familiarity: While models adapt immediately, include a brief primer for human users if needed.
  • Fact Updates: Maintain validity windows in the graph to avoid serving outdated information.

The system is robust even for users without strong visualization skills, as retrieval is database-driven.

Conclusion

MemPalace represents a significant advancement in AI memory management by combining ancient spatial memory principles with modern local-first technology. It delivers unmatched recall accuracy, complete privacy, and zero ongoing costs.

Explore the project today on the official GitHub repository at github.com/milla-jovovich/mempalace. Install MemPalace, build your first virtual palace, and give your AI the persistent memory it has long needed. The future of reliable, private AI context starts here.

Share this article