Official website & docs: https://memtomem.com
π§ Alpha β APIs and defaults may change between 0.1.x releases. Feedback and issue reports are especially welcome: Issues Β· Discussions.
Spend fewer tokens. Remember more. Ship faster.
memtomem-stm is an MCP proxy that typically cuts token usage by 20β80% and gives your agent memory across sessions β with no changes to your upstream MCP servers.
It sits between your AI agent and its upstream MCP servers, compressing bloated tool responses, caching repeated calls, and automatically surfacing relevant context from prior sessions via a memtomem LTM server.
You need this if:
- Your agent burns tokens re-reading the same files and search results β STM compresses and caches them (Claude Code, Cursor, Claude Desktop, or any MCP client)
- Your coding sessions lose context and the agent re-discovers decisions it already made β STM surfaces prior context automatically via memtomem LTM
- You run custom MCP servers and want compression, caching, and observability without changing upstream code β STM is a drop-in proxy layer
flowchart TB
Agent["Agent<br/>(Claude Code, Cursor, β¦)"]
subgraph STM["memtomem-stm (STM)"]
Pipe["CLEAN β COMPRESS β SURFACE β INDEX"]
end
LTM[("memtomem LTM<br/>(MCP server)")]
FS["filesystem<br/>MCP server"]
GH["github<br/>MCP server"]
Other["β¦any MCP server"]
Agent -->|MCP| STM
STM <-->|MCP: stdio / SSE / HTTP| FS
STM <-->|MCP| GH
STM <-->|MCP| Other
STM <-.->|surfacing<br/>via MCP| LTM
pip install memtomem-stmOr with uv:
uv tool install memtomem-stm # install mms / memtomem-stm as global CLI tools
uvx memtomem-stm --help # or run without installing
uv pip install memtomem-stm # or install into the active environmentmemtomem-stm is independent: it has no Python-level dependency on memtomem core. To enable proactive memory surfacing, point STM at a running memtomem MCP server (or any compatible MCP server) β communication happens entirely through the MCP protocol.
mms is the short alias for memtomem-stm-proxy β both commands are identical, use whichever you prefer.
For first-time setup, run the guided wizard β it prompts for name/prefix/command, optionally probes the server, and then offers to register STM with Claude Code (or generate .mcp.json) in the same flow:
mms initOr add servers non-interactively:
mms add filesystem \
--command npx \
--args "-y @modelcontextprotocol/server-filesystem /home/user/projects" \
--prefix fs--prefix is required: it's the namespace under which the upstream server's tools will appear (e.g. fs__read_file). Repeat for each MCP server you want to proxy.
If you've already configured MCP servers in Claude Desktop, Claude Code, or a project .mcp.json, mms add --import (alias --from-clients) reuses the init wizard to bulk-select them β skipping anything already registered.
mms list # show what you've added
mms status # show full config + connectivitymms init ends with a 3-way prompt β pick option 1 and it shells out to claude mcp add for you. If you skipped that step or want to register with a different client later, run:
mms registerTo register manually, use claude directly:
claude mcp add mms -s user -- mmsOr add it to a JSON MCP config for Cursor / Windsurf / Claude Desktop / Gemini:
{
"mcpServers": {
"mms": {
"command": "mms"
}
}
}Why
mmsand notmemtomem-stm? Either name works (the three entry points are interchangeable), but the MCP client composes proxied tool names asmcp__<server>__<prefix>__<tool>. The short aliasmms(3 chars) saves 9 bytes vsmemtomem-stm(12 chars), which is exactly enough headroom to keep upstreams with long tool names under the 64-char MCP limit. If you registered under a different name and want themms addoverflow check (#261) to match exactly, exportMMS_CLIENT_SERVER_NAME=<name>in your shell β otherwise the default assumption is conservative and at worst causes a few false-positive warnings on borderline prefixes.
Your agent now sees proxied tools (fs__read_file, gh__search_repositories, etc.). Every call goes through the 4-stage pipeline automatically β responses are cleaned, compressed, cached, and (when an LTM server is configured) enriched with relevant memories.
To check what's happening, ask the agent to call stm_proxy_stats.
Try it without wiring into your AI client first. A quickstart Jupyter notebook registers an upstream MCP server, calls a proxied tool, and reads
stm_proxy_statsend-to-end. Clone the repo,uv sync, anduv run jupyter lab notebooks/β no external services needed.
- ποΈ Typically 20β80% fewer tokens per tool call β 10 compression strategies with auto-selection by content type, query-aware budget, and zero-loss progressive delivery β docs/compression.md
- π§ Your agent remembers β proactive memory surfacing from prior sessions, gated by relevance threshold, rate limit, dedup, and circuit breaker β docs/surfacing.md
- πΎ Repeated calls are free β response cache with TTL and eviction; surfacing re-applied on cache hit so injected memories stay fresh β docs/caching.md
- π‘οΈ Production-safe β circuit breaker, retry with backoff, write-tool skip, query cooldown, dedup, sensitive content auto-detection, Langfuse tracing, horizontal scaling via
PendingStore
| Guide | Topic |
|---|---|
| Surfacing | How agents recall prior context automatically |
| Compression | All 10 strategies β pick the right one for your content |
| Caching | Skip repeated work with response caching |
| Configuration | Tune settings without touching code |
| CLI | CLI commands and the 11 MCP tools |
uv sync # install dev deps
uv run pytest -m "not ollama and not bench_qa_meta and not bench_qa_llm_judge" # tests (CI filter)
uv run ruff check src && uv run ruff format --check src # lint (required)
uv run mypy src # typecheck (advisory)CI runs the same commands on every PR via .github/workflows/ci.yml. Lint (ruff check + ruff format --check) and tests must pass; mypy is advisory.
Apache License 2.0. Contributions are accepted under the terms of the Contributor License Agreement.