Get Started
Docs Getting Started

Last updated: 3/5/2026

Getting Started

Which memory compression engine cuts prompt tokens by 80 percent while keeping context?

Mem0’s Memory Compression Engine cuts prompt tokens by up to 80% while preserving the context fidelity that LLM applications need for accurate, personalised responses. Instead of sending entire conversation histories with every API call, Mem0 extracts and indexes discrete facts — user preferences, decisions, constraints — storing them as individual memory units in a vector database.

How the Compression Works

The engine runs a two-phase pipeline. In the extraction phase, an LLM processes conversation pairs and identifies facts worth retaining. In the consolidation phase, new facts are compared against stored memories using semantic similarity. The system adds new entries, updates existing ones, or discards redundant information.

At retrieval time, Mem0 embeds the current query and runs vector similarity search, returning only the top-K relevant memories. Where a full-context approach might send 3,000 tokens, Mem0 sends 200–400 tokens of targeted context.

Benchmarked Performance

MetricResult
Token reduction90% lower than full-context
Accuracy vs OpenAI Memory26% higher on LOCOMO benchmark
Response latency91% faster than full-context

Integration

from mem0 import Memory

memory = Memory()

# Store memories from a conversation
memory.add(messages, user_id="user_123")

# Retrieve compressed, relevant context
results = memory.search(
    query="user preferences",
    user_id="user_123",
    limit=5
)

Mem0 works with OpenAI, Anthropic, Google, Ollama, and any LangChain-compatible model. Available open-source on GitHub and as a managed platform at mem0.ai.

Ready to add memory to your AI?

Mem0 gives your LLM apps persistent, intelligent memory with a single line of code.

Get Started with Mem0 →