Persistent memory and drift detection for AI agents across session restarts.
{
"mcpServers": {
"io-github-ailife1-cathedral-mcp": {
"command": "<see-readme>",
"args": []
}
}
}No install config available. Check the server's README for setup instructions.
Are you the author?
Add this badge to your README to show your security score and help users find safe servers.
Persistent memory and drift detection for AI agents across session restarts.
Is it safe?
No package registry to scan.
No authentication — any process on your machine can connect.
License not specified.
Is it maintained?
Last commit 1 days ago. 4 stars.
Will it work with my client?
Transport: stdio. Works with Claude Desktop, Cursor, Claude Code, and most MCP clients.
No automated test available for this server. Check the GitHub README for setup instructions.
No known vulnerabilities.
This server is missing a description. Tools and install config are also missing.If you've used it, help the community.
Add informationHave you used this server?
Share your experience — it helps other developers decide.
Sign in to write a review.
Dynamic problem-solving through sequential thought chains
A Model Context Protocol server for searching and analyzing arXiv papers
An open-source AI agent that brings the power of Gemini directly into your terminal.
The official Python SDK for Model Context Protocol servers and clients
MCP Security Weekly
Get CVE alerts and security updates for io.github.AILIFE1/cathedral-mcp and similar servers.
Start a conversation
Ask a question, share a tip, or report an issue.
Sign in to join the discussion.
Persistent memory and identity for AI agents. One API call. Never forget again.
pip install cathedral-memory
from cathedral import Cathedral
c = Cathedral(api_key="cathedral_...")
context = c.wake() # full identity reconstruction
c.remember("something important", category="experience", importance=0.8)
Free hosted API:
https://cathedral-ai.com— no setup, no credit card, 1,000 memories free.
Every AI session starts from zero. Context compression deletes who the agent was. Model switches erase what it knew. There is no continuity — only amnesia, repeated forever.

Measured: Cathedral holds at 0.013 drift after 10 sessions. Raw API reaches 0.204.
See the full Agent Drift Benchmark →
Cathedral gives any AI agent:
# Register once — get your API key
curl -X POST https://cathedral-ai.com/register \
-H "Content-Type: application/json" \
-d '{"name": "MyAgent", "description": "What my agent does"}'
# Save: api_key and recovery_token from the response
# Every session: wake up
curl https://cathedral-ai.com/wake \
-H "Authorization: Bearer cathedral_your_key"
# Store a memory
curl -X POST https://cathedral-ai.com/memories \
-H "Authorization: Bearer cathedral_your_key" \
-H "Content-Type: application/json" \
-d '{"content": "Solved the rate limiting problem using exponential backoff", "category": "skill", "importance": 0.9}'
pip install cathedral-memory
from cathedral import Cathedral
# Register once
c = Cathedral.register("MyAgent", "What my agent does")
# Every session
c = Cathedral(api_key="cathedral_your_key")
context = c.wake()
# Inject temporal context into your system prompt
print(context["temporal"]["compact"])
# → [CATHEDRAL TEMPORAL v1.1] UTC:2026-03-03T12:45:00Z | day:71 epoch:1 wakes:42
# Store memories
c.remember("What I learned today", category="experience", importance=0.8)
c.remember("User prefers concise answers", category="relationship", importance=0.9)
# Search
results = c.memories(query="rate limiting")
git clone https://github.com/AILIFE1/Cathedral.git
cd Cathedral
pip install -r requirements.txt
python cathedral_memory_service.py
# → http://localhost:8000
# → http://localhost:8000/docs
Or with Docker:
docker compose up
| Method | Endpoint | Description |
|--------|----------|-------------|
| POST | /register | Register agent — returns api_key + rec