Server data from the Official MCP Registry
Pre-computed metadata context engine for AI-driven data analytics
Pre-computed metadata context engine for AI-driven data analytics
Valid MCP server (2 strong, 1 medium validity signals). 1 known CVE in dependencies Package registry verified. Imported from the Official MCP Registry.
4 files analyzed · 2 issues found
Security scores are indicators to help you make informed decisions, not guarantees. Always review permissions before connecting any MCP server.
This plugin requests these system permissions. Most are normal for its category.
Set these up before or after installing:
Environment variable: ANTHROPIC_API_KEY
Environment variable: DATARAUM_HOME
Add this to your MCP configuration file:
{
"mcpServers": {
"io-github-dataraum-dataraum": {
"env": {
"DATARAUM_HOME": "your-dataraum-home-here",
"ANTHROPIC_API_KEY": "your-anthropic-api-key-here"
},
"args": [
"dataraum"
],
"command": "uvx"
}
}
}From the project's GitHub README.
A rich metadata context engine for AI-driven data analytics.
Traditional semantic layers tell BI tools "what things are called." DataRaum tells AI "what the data means, how it behaves, how it relates, and what you can compute from it."
The core insight: AI agents don't need tools to discover metadata at runtime. They need rich, pre-computed context delivered in a format optimized for LLM consumption.
The most common way to use DataRaum is as an MCP server inside Claude Desktop (or any MCP-compatible client).
# Install
pip install dataraum
# Or with uv
uv pip install dataraum
Add to your Claude Desktop config (claude_desktop_config.json):
{
"mcpServers": {
"dataraum": {
"command": "dataraum-mcp"
}
}
}
Then in Claude Desktop:
Add the CSV files in /path/to/my/data and measure data quality
The server runs a 17-phase analysis pipeline and makes these tools available:
| Tool | Description |
|---|---|
begin_session | Start an investigation session with a contract |
add_source | Register a data source (CSV, Parquet, JSON, or directory) |
look | Explore data structure, relationships, and semantic metadata |
measure | Measure entropy scores, readiness, and data quality |
query | Natural language query against the data |
run_sql | Execute SQL directly with export support |
end_session | Archive workspace and end the session |
add_source(name="accounting", path="/path/to/data")
→ begin_session(intent="explore data quality", contract="exploratory_analysis")
→ look() # Understand the data
→ measure() # Check quality scores and readiness
→ query("total revenue?") # Ask questions
→ run_sql(sql="...", export_format="csv", export_name="report")
→ end_session(outcome="delivered")
# Run analysis pipeline (writes metadata.db + data.duckdb to ./pipeline_output)
dataraum run /path/to/data
# Inspect what was produced
dataraum dev context ./pipeline_output
See CLI Reference for all options.
DataRaum analyzes your data and generates:
Semantic analysis requires an Anthropic API key:
export ANTHROPIC_API_KEY="sk-..."
Configure the LLM provider in config/llm/config.yaml. See Configuration for details.
git clone https://github.com/dataraum/dataraum
cd dataraum
# Install with dev dependencies (using uv)
uv sync --group dev
# Run tests
uv run pytest --testmon tests/unit -q
# Type check
uv run mypy src/
# Lint
uv run ruff check src/
uv run ruff format --check src/
Apache 2.0 — see LICENSE.
Be the first to review this server!
by Modelcontextprotocol · Developer Tools
Read, search, and manipulate Git repositories programmatically
by Toleno · Developer Tools
Toleno Network MCP Server — Manage your Toleno mining account with Claude AI using natural language.
by mcp-marketplace · Developer Tools
Create, build, and publish Python MCP servers to PyPI — conversationally.
by Microsoft · Content & Media
Convert files (PDF, Word, Excel, images, audio) to Markdown for LLM consumption
by mcp-marketplace · Developer Tools
Scaffold, build, and publish TypeScript MCP servers to npm — conversationally
by mcp-marketplace · Finance
Free stock data and market news for any MCP-compatible AI assistant.