Self-hosted AI memory · Apache-2.0
Your data stays yours.
Memory that runs on your machine. Drop-in for Claude.
Core capabilities
Memory that works like yours
Three primitives. Everything your agent needs to learn, recall, and comply.
Remember
Every conversation, preference, and decision — persisted automatically. MemDB extracts entities, classifies memory types, and deduplicates. Your agent gets smarter with every interaction.
Recall
Ask in natural language, get the right context back. Hybrid vector + fulltext search with temporal decay — recent memories rank higher. Results in under 50ms.
Forget
GDPR compliance built in. Delete a single memory, a user's entire history, or everything in a namespace. Instant. No reindex. No residual embeddings.
Three Claude integration surfaces
Plug MemDB into Claude. Pick your surface.
Same memory backend, three install paths. The Anthropic memory_20250818 adapter is a drop-in — no code changes for users already on the official tool spec.
Claude Code plugin
Auto-injects memories into your IDE context. Install via the Claude Code plugin manager and forget it exists.
MCP server
Built-in MCP at :8001 for Claude Desktop, Cursor, Windsurf, or any MCP-compatible host. stdio + HTTP transport.
Anthropic memory_20250818 adapter
Python drop-in for the official Claude memory tool. Swap one import — no SDK changes, no prompt rewrites.
Benchmarks
| System | LoCoMo LLM Judge | License |
|---|---|---|
| 1. Memobase | 75.78% | Commercial cloud |
| 2. Zep | 75.14% | Apache-2.0 + Cloud |
| 3. MemOS | 73.31% | Apache-2.0 |
| MemDB v0.23.0 | 72.50% | Apache-2.0 |
| 5. Mem0 | 66.88% | Commercial cloud |
LoCoMo chat-50 stratified, LLM Judge by Gemini Flash, category 5 excluded. Methodology and raw results: MILESTONES.md.
Start in minutes
# Add a memory
curl -X POST https://api.memdb.ai/product/add \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"user_id": "agent-123",
"messages": [{"role": "user", "content": "I prefer bullet points"}]
}'
# Search memories
curl -X POST https://api.memdb.ai/product/search \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{"user_id": "agent-123", "query": "formatting preferences", "top_k": 5}'
# Delete memories
curl -X POST https://api.memdb.ai/product/delete_memory \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{"user_id": "agent-123", "memory_ids": ["mem_abc"]}'Built to run offline
No External APIs
ONNX embeddings run locally. LLM extractor works with Ollama, vLLM, or any OpenAI-compatible API — no cloud LLM required. Fully air-gappable.
One Database
Everything in PostgreSQL: vectors (pgvector HNSW), graph (Apache AGE), fulltext (tsvector). No separate graph DB, no managed vector service. One backup, one migration path.
MCP Native
Built-in MCP server with stdio + HTTP transport. Claude Code, Cursor, Windsurf — any MCP client connects instantly. No SDK integration needed. Your agent gets memory in one config line.
Frequently asked questions
Join the waitlist
Be first to know when the hosted API opens. Self-hosted is available now on GitHub.