Connect with us

Thought Leaders

Milla Jovovich’s MemPalace Aims to Solve AI’s Memory Problem

mm
A woman standing in a luminous, neoclassical hallway, interacting with glowing holographic data structures representing a

Millions of people open a chat window daily and start explaining themselves to artificial intelligence (AI). It listens attentively, instantly generates a clever-sounding answer, and then, when the session ends, forgets every single detail about the interaction.

The scale of this ritual is staggering. ChatGPT alone handles over one billion queries per day, with more than 800 million weekly active users as of late 2025. Generative AI adoption has reached over 16% of the world’s population, a number that didn’t exist meaningfully three years ago.

An enormous infrastructure with an increasing environmental price supports this model: U.S. data centers consumed 183 terawatt-hours of electricity in 2024, more than 4% of the country’s total usage, or roughly equal to Pakistan’s annual electricity demand.

Due to the lack of memory of AI systems, a big percentage of that energy is spent re-establishing context. Repeated explanations, project reintroductions, and context dumps at the start of every session are wasted computation.

Memory Is What Turns a Tool Into a Collaborator

AI assistants have no persistent memory by default. This wouldn’t matter if we used AI like a calculator: punch in a number, get a result, move on.

But most people don’t use it that way anymore. They have long, iterative, deeply contextual conversations with AI—building things over weeks or months, developing shared language, decisions, and history. The amount of context the AI can actively keep in mind at any given time can vary depending on the subscription tier.

So far, AI has proven to be a wonderful tool, but since early development stages it has aimed towards being considered a companion. That ambition requires memory. Without it, progress will continue to reset.

Persistent memory changes what AI can do in practice. A developer gets an AI that retains architectural decisions and the reasoning behind them. A team gets one that knows the project history without being re-briefed. A writer gets one that has accumulated knowledge about their work over time. The capability of the model matters less than whether it can actually accumulate knowledge about the person using it.

Why This Has Been Hard to Solve

The challenge isn’t just storage but retrieval. In theory, you can feed every past conversation into a new session. But that quickly becomes computationally absurd. Context windows, though expanding, aren’t infinite. Dumping months of unstructured chat into a prompt is not only ineffective but also time and energy consuming.

Paras Pandey, a data engineer, puts the core difficulty plainly: “AI memory is really a retrieval fidelity problem dressed up as a storage problem. You can persist anything, the hard part is retrieving the right slice of it at inference time without hallucinating the gaps. That’s a harder version of what we’ve been solving in data systems for years, and the field is still early.”

Current AI memory approaches involve letting systems decide what’s worth remembering. But letting AI decide what matters often throws away exactly the kind of nuanced context that made the original exchange valuable. You keep the general idea but lose the entire conversation where you explained your specific concerns,and the alternatives you considered and rejected.

The ideal scenario would be to make the right information findable at the right moment.

Enter MemPalace

This is precisely the problem that MemPalace, a recently released open-source project, takes aim at. Rather than summarizing or discarding, it stores conversations in full and builds a navigable structure around them, borrowing from the ancient Greek technique of the memory palace, where orators would mentally place ideas in specific rooms of an imagined building to recall them later.

What makes MemPalace notable isn’t just the elegance of the approach. It’s the results. In standard academic benchmarks for AI memory retrieval, MemPalace has posted the highest scores ever published for a free system, and it does so while running entirely on your own machine, with no subscription, no cloud dependency, and no external API required.

Competing commercial services charge anywhere from $20 to $250 a month for comparable, and often worse-performing functionality.

That combination of best-in-class performance, fully local, and completely free is unusual enough to be worth paying attention to. And because it runs on your hardware rather than remote servers, every query you route through MemPalace is one that doesn’t add to the ballooning energy ledger of the data center industry.

The Bigger Picture

MemPalace is one project, but it points to something larger: recognition that persistent memory is not a premium add-on to AI systems, It’s a fundamental feat for the new use cases of AI.

The project was built by a small team, Milla Jovovich (Yes, the actress from Resident Evil), Ben Sigman, and Claude, and is listed as having only seven commits.

That a system outperforming commercial products with dedicated engineering teams came from such a lean effort says something about where the real difficulty lies.

The problem wasn’t computers or resources. It was a clearer model of what memory actually needs to do.

Juan Pablo Aguirre Osorio is a contributing reporter to Espacio Media Incubator. With a background in full-stack engineering, Juan Pablo brings a technical background to his reporting on cutting-edge technologies, including AI. His work has been featured in HackerNoon, The Sociable, and others, and he was previously a Student Ambassador at Microsoft.