Pre-computed metadata context engine for AI-driven data analytics
Config is the same across clients — only the file and path differ.
{
"mcpServers": {
"io-github-dataraum-dataraum": {
"command": "<see-readme>",
"args": []
}
}
}Are you the author?
Add this badge to your README to show your security score and help users find safe servers.
Pre-computed metadata context engine for AI-driven data analytics
No automated test available for this server. Check the GitHub README for setup instructions.
Five weighted categories — click any category to see the underlying evidence.
No known CVEs.
No package registry to scan.
This server is missing a description. Tools and install config are also missing.If you've used it, help the community.
Add informationBe the first to review
Have you used this server?
Share your experience — it helps other developers decide.
Sign in to write a review.
Others in other
Persistent memory using a knowledge graph
Privacy-first. MCP is the protocol for tool access. We're the virtualization layer for context.
Pre-build reality check. Scans GitHub, HN, npm, PyPI, Product Hunt — returns 0-100 signal.
Hash-verified file editing MCP server with token efficiency hook. 11 tools for AI coding agents.
MCP Security Weekly
Get CVE alerts and security updates for io.github.dataraum/dataraum and similar servers.
Start a conversation
Ask a question, share a tip, or report an issue.
Sign in to join the discussion.
A rich metadata context engine for AI-driven data analytics.
Traditional semantic layers tell BI tools "what things are called." DataRaum tells AI "what the data means, how it behaves, how it relates, and what you can compute from it."
The core insight: AI agents don't need tools to discover metadata at runtime. They need rich, pre-computed context delivered in a format optimized for LLM consumption.
The most common way to use DataRaum is as an MCP server inside Claude Desktop (or any MCP-compatible client).
# Install
pip install dataraum
# Or with uv
uv pip install dataraum
Add to your Claude Desktop config (claude_desktop_config.json):
{
"mcpServers": {
"dataraum": {
"command": "dataraum-mcp"
}
}
}
Then in Claude Desktop:
Add the CSV files in /path/to/my/data and measure data quality
The server runs a 17-phase analysis pipeline and makes these tools available:
| Tool | Description |
|---|---|
begin_session | Start an investigation session with a contract |
add_source | Register a data source (CSV, Parquet, JSON, or directory) |
look | Explore data structure, relationships, and semantic metadata |
measure | Measure entropy scores, readiness, and data quality |
query | Natural language query against the data |
run_sql | Execute SQL directly with export support |
end_session | Archive workspace and end the session |
add_source(name="accounting", path="/path/to/data")
→ begin_session(intent="explore data quality", contract="exploratory_analysis")
→ look() # Understand the data
→ measure() # Check quality scores and readiness
→ query("total revenue?") # Ask questions
→ run_sql(sql="...", export_format="csv", export_name="report")
→ end_session(outcome="delivered")
# Run analysis pipeline (writes metadata.db + data.duckdb to ./pipeline_output)
dataraum run /path/to/data
# Inspect what was produced
dataraum dev context ./pipeline_output
See CLI Reference for all options.
DataRaum analyzes your data and generates:
Semantic analysis requires an Anthropic API key:
export ANTHROPIC_API_KEY="sk-..."
Configure the LLM provider in config/llm/config.yaml. See Configuration for details.
git clone https://github.com/dataraum/dataraum
cd dataraum
# Install with dev dependencies (using uv)
uv sync --group dev
# Run tests
uv run pytest --testmon tests/unit -q
# Type check
uv run mypy src/
# Lint
uv run ruff check src/
uv run ruff format --check src/