Server data from the Official MCP Registry
Score your specs before feeding them to an LLM. MCP server with radar chart visualization.
Score your specs before feeding them to an LLM. MCP server with radar chart visualization.
Valid MCP server (2 strong, 4 medium validity signals). No known CVEs in dependencies. Imported from the Official MCP Registry. 1 finding(s) downgraded by scanner intelligence.
5 files analyzed · 1 issue found
Security scores are indicators to help you make informed decisions, not guarantees. Always review permissions before connecting any MCP server.
Add this to your MCP configuration file:
{
"mcpServers": {
"io-github-openpoem-spec-score-mcp": {
"args": [
"-y",
"spec-score-mcp"
],
"command": "npx"
}
}
}From the project's GitHub README.
Score your specs before Claude builds from them.
A balanced spec produces balanced code. An unbalanced spec produces creative fiction.
When your spec is detailed on some axes but vague on others, Claude doesn't ask for clarification — it fills in the blanks. The result compiles, the tests pass, but it's not what you meant.
This tool catches that before you start building. It scores your spec on 4 axes, tells you which one is weakest, and gives you a concrete tip to fix it.
| Axis | Question it answers |
|---|---|
| completeness | Can Claude understand the full scope of what to build? |
| clarity | Is there only one way to interpret this spec? |
| constraints | Does Claude know what NOT to build? |
| specificity | Are there concrete, testable details? |
Each axis scores 0.0 to 1.0. The balance score measures how evenly the 4 axes are covered.
Balance matters more than individual scores. A spec scoring 0.50 on all 4 axes (balance: 0.97) will produce better output than one scoring 0.95 / 0.95 / 0.20 / 0.90 (balance: 0.58). Why? That one weak axis — constraints at 0.20 — is exactly where Claude will improvise. You described what to build in detail, but forgot to say what's out of scope. So Claude builds everything you asked for, plus features you didn't.
On the radar chart: an even diamond beats a sharp spike.
| Verdict | What it means |
|---|---|
| SHIP IT | Spec is ready — Claude knows what to build and what not to |
| ALMOST | One axis needs a small fix before you start |
| DRAFT | Multiple axes need work, but the structure is there |
| VAGUE | Well-organized but too abstract to act on |
| UNBOUNDED | Clear goal but no boundaries — Claude will over-build |
| OVER-CONSTRAINED | Lots of rules but unclear what the actual goal is |
| SKETCH | Starting point — needs detail on most axes |
Not SHIP IT yet? The tool tells you which axis is weakest and what to add. Fix that axis, re-score, repeat. Most specs reach SHIP IT in 2-3 rounds.
git clone https://github.com/openpoem/spec-score-mcp.git
cd spec-score-mcp && npm install && npm run build
claude mcp add spec-score -- node $(pwd)/dist/mcp.js
The 3 tools are now available in every Claude Code session.
Clone this repo to get the built-in slash commands:
/project:scan my-feature-spec.md
Reads the file, scores it, and writes a my-feature-spec.md.scored.md with scores, verdict, tip, and a radar chart.
/project:compare blueprint.md implementation.md
Scores both files and writes a compared.scored.md with side-by-side radar charts.
The 3 MCP tools work in any Claude Code conversation:
| Tool | What it does |
|---|---|
spec_score | Scores a spec on 4 axes, returns balance score and verdict |
spec_visualize | Generates an SVG radar chart from scores |
spec_compare | Side-by-side comparison of two scored specs |
Ask Claude: "Score this spec", "Show me the radar chart", or "Compare these two specs".
This tool scoring its own spec — four rounds, each fixing the weakest axis:
Build a spec scoring tool
UNBOUNDED 0.12 Tip: What does 'scoring' mean? What axes? What output?
One axis is high (clarity — the goal is clear), everything else is near zero. Claude would build... anything. A web app? A CLI? A VS Code extension? No way to know.
Build an MCP server that scores specs on 4 axes: completeness, clarity, constraints, specificity. Each axis is 0.0-1.0. Returns a balance score and verdict.
ALMOST 0.67 Tip: What are the verdicts? What does the tool NOT do?
Now Claude knows what to build. But constraints are still weak — it might add auto-fixing, CI integration, a database.
Three tools: spec_score, spec_visualize, spec_compare. Non-goals: no auto-fixing, no CI integration, no storage.
SHIP IT 0.84 Tip: Add testable criteria — what balance maps to which verdict?
Crossed the threshold. Claude now knows what to build AND what not to build. Specificity is still the weakest axis.
Balance = 1 - sqrt(variance)/mean. SHIP IT > 0.75, ALMOST > 0.60, plus pattern-based verdicts. Node.js, MCP SDK, stdio transport.
SHIP IT 0.95 Spec is ready for implementation.
Four rounds: 0.12 → 0.67 → 0.84 → 0.95. Each round fixed exactly one thing.
v / ||v||1 - sqrt(variance) / meanThe scoring intelligence comes from Claude, not from the algorithm. The algorithm only measures balance.
src/
mcp.ts # MCP server (3 tools)
score.ts # Scoring engine
visualize.ts # SVG radar charts
.claude/
commands/
scan.md # /project:scan command
compare.md # /project:compare command
OpenPoem — spec-score-mcp
MIT License.
© 2026 OpenPoem. info@openpoem.org
Be the first to review this server!
by Modelcontextprotocol · Developer Tools
Read, search, and manipulate Git repositories programmatically
by Toleno · Developer Tools
Toleno Network MCP Server — Manage your Toleno mining account with Claude AI using natural language.
by mcp-marketplace · Developer Tools
Create, build, and publish Python MCP servers to PyPI — conversationally.
by Microsoft · Content & Media
Convert files (PDF, Word, Excel, images, audio) to Markdown for LLM consumption
by mcp-marketplace · Developer Tools
Scaffold, build, and publish TypeScript MCP servers to npm — conversationally
by mcp-marketplace · Finance
Free stock data and market news for any MCP-compatible AI assistant.