Persistent memory for Claude Code and Cursor. Stop re-explaining your project every session.
{
"mcpServers": {
"io-github-sumanaddanki-suma-memory": {
"command": "<see-readme>",
"args": []
}
}
}No install config available. Check the server's README for setup instructions.
Are you the author?
Add this badge to your README to show your security score and help users find safe servers.
Persistent memory for Claude Code and Cursor. Stop re-explaining your project every session.
Is it safe?
No package registry to scan.
No authentication — any process on your machine can connect.
License not specified.
Is it maintained?
Last commit 3 days ago.
Will it work with my client?
Transport: . Compatibility not confirmed.
Stop re-explaining your project to Claude every time you start a new chat.
Your repos now have permanent memory. SUMA gives any MCP-compatible AI client (Claude Code, Cursor, Devin) a persistent knowledge graph that remembers architectural decisions, bug root causes, and project rules — across sessions, across machines, across your entire team.
Get an API key at sumapro.quadframe.work — free tier available.
Add to your .mcp.json:
{
"mcpServers": {
"suma-memory": {
"url": "https://sumapro.quadframe.work/mcp",
"headers": {
"Authorization": "Bearer sk_live_your_key_here"
}
}
}
}
That's it. No local server. No Docker. No npm install. SUMA runs on Cloud Run — stateless, auto-scaled, always available.
After installing, run this once per repo to seed your permanent context:
suma_ingest(text="Project: [name]. Framework: [Next.js / Flask / etc].
Auth lives in: [path/to/auth.py]. Database: [PostgreSQL / SQLite / etc].
Rules never to break: [e.g. never store plaintext keys, all routes require org_id filter].
Deployment target: [Cloud Run / Vercel / etc].")
From this point forward, every new session inherits this context. You never explain it again.
SUMA stores knowledge in a weighted graph. Every node has a gravity score across four dimensions:
When you call suma_search, the K-WIL gravity algorithm traverses the graph and returns the highest-relevance context — not a flat list of chunks, not a raw embedding match, but the facts that actually matter for what you're doing right now.
| Tool | What it does |
|---|---|
suma_ping | Health check — verify connection and API key |
suma_ingest | Add knowledge to the graph (architecture decisions, bug fixes, rules) |
suma_search | Retrieve relevant context by natural language query |
suma_talk | Search + learn in one call — retrieves context and updates graph |
suma_correct | Fix wrong information — supersedes original, queues replacement |
suma_clean | Remove noise nodes that pollute search results |
# After finalizing a decision:
suma_ingest(text="We chose REST over GraphQL. Root cause: GraphQL N+1 queries
caused 3x latency on /search. Architect ruling Apr 10 2026.")
# Next session, cold start — full context in one call:
suma_search(query="why did we switch to REST?")
# → Returns ruling with full context. No re-explaining.
# After fixing a hard bug:
suma_ingest(text="Cloud Run WebSocket bug: asyncio.run() in daemon thread killed
by Cloud Run recycling. Fix: use asyncio.get_event_loop() instead.
Never use asyncio.run() in long-lived Cloud Run services.")
# Six months later, same error:
suma_search(query="asyncio cloud run daemon thread crash")
# → Root cause retrieved instantly. Hours saved.
Architect, developer, and QA agents each write to SUMA using their own sessions. Their knowledge merges into one shared org graph. When QA asks "what did the architect decide about auth?", it retrieves the architect's ruling — zero explicit handoff required.
Anti-flood protection: Each source machine is rate-limited to 5 ingests per 60 seconds. Runaway agent loops are broken gracefully — the 6th request returns {"status": "throttled"} without crashing or corrupting the graph.
Multi-tenant isolation: Every node is scoped to org_id at the database layer. Two organizations
No automated test available for this server. Check the GitHub README for setup instructions.
This server is missing a description. Tools and install config are also missing.If you've used it, help the community.
Add informationNo known vulnerabilities.
Have you used this server?
Share your experience — it helps other developers decide.
Sign in to write a review.
Dynamic problem-solving through sequential thought chains
A Model Context Protocol (MCP) server and CLI that provides tools for agent use when working on iOS and macOS projects.
The official Python SDK for Model Context Protocol servers and clients
An open-source AI agent that brings the power of Gemini directly into your terminal.
MCP Security Weekly
Get CVE alerts and security updates for io.github.sumanaddanki/suma-memory and similar servers.
Start a conversation
Ask a question, share a tip, or report an issue.
Sign in to join the discussion.