Skip to main content
SubstratiaSubstratia
Start HereToolsReviewsResearchBlogDocsGitHub

Stay updated on agent memory infrastructure

New tools, memory patterns, and research on persistent AI identity.

SubstratiaSubstratiaMemory Infrastructure for AI Agents
Start HereToolsReviewsResearchBlogDocsGitHub

Intelligence is substrate-agnostic.

Built by practitioners.

Privacy PolicyTerms of Service
← Back to Blog
Releasememory-mcpv2.5

What's New in memory-mcp v2.5: From Python to TypeScript

January 11, 20266 min read

If you found memory-mcp through an older article talking about Python, sentence transformers, and vector similarity search - that's the old version. Here's what changed and why.

TL;DR

memory-mcp v2.5 is a complete rewrite: Python → TypeScript, embeddings → FTS5, pip → npx. It's simpler, faster, and has zero external dependencies.

The Old Version (Pre-v2.0)

The original memory-mcp was built in Python with:

  • Sentence transformers for embeddings
  • Vector similarity search
  • Python 3.8-3.12 required
  • pip install workflow
  • Heavy dependencies (PyTorch, transformers)

It worked, but had problems: slow startup, large memory footprint, complex installation, and embedding models that needed downloading.

The New Version (v2.5)

We completely rewrote memory-mcp with a different philosophy: simplicity over features.

AspectOld (Python)New (v2.5)
LanguagePython 3.8+TypeScript
SearchVector embeddingsFTS5 full-text
StorageJSON filesSQLite
Installpip + requirementsnpx (one command)
DependenciesPyTorch, transformersMCP SDK + better-sqlite3
Startup30+ seconds<1 second
Embedding modelRequired (500MB+)None needed

Why We Dropped Embeddings

Embeddings are powerful but overkill for most memory use cases. Here's why:

  • Memory isn't search. You're not searching millions of documents - you're recalling dozens to hundreds of memories.
  • Keywords work. FTS5 with BM25 ranking is fast and accurate for reasonable memory sizes.
  • Simplicity wins. No model downloads, no GPU, no cold start.

For most users, "find memories about authentication" works perfectly with keyword search. If you truly need semantic search at scale, use a dedicated solution like Qdrant or Pinecone.

New Features in v2.5

  • Token budgeting - Recall respects your context limits
  • Auto-summarization - Long memories get summarized on store
  • Entity extraction - Automatic tagging of people, projects, concepts
  • Soft deletes - Audit trail preserved for debugging
  • Hybrid relevance - Scoring combines recency + importance + frequency

Migration Guide

If you were using the old Python version:

  1. Export your memories from the old JSON files (optional - start fresh is fine)
  2. Uninstall the old Python package
  3. Install the new version:
npx @whenmoon-afk/memory-mcp

Or add to your Claude Desktop config:

{
  "mcpServers": {
    "memory": {
      "command": "npx",
      "args": ["-y", "@whenmoon-afk/memory-mcp"]
    }
  }
}

Restart Claude Desktop. You now have three tools: memory_store, memory_recall, and memory_forget.

What's Next

We're continuing to improve memory-mcp with:

  • Better search relevance and ranking
  • Memory dashboard to view/edit what AI remembers
  • Performance optimizations for large memory stores
  • Improved integration with Claude Code workflows

End the Amnesia Loop

Give your agent persistent memory. Install in seconds, no Python required.

View on GitHub
S
Substratia Team
Building developer tools for Claude Code

Related Posts

MCP Memory Servers Compared
memory-mcp vs the alternatives
How to Build Claude Agents
A complete guide to CLAUDE.md files