Server data from the Official MCP Registry
Generate MCP servers from natural language, OpenAPI specs, or database schemas.
Generate MCP servers from natural language, OpenAPI specs, or database schemas.
Valid MCP server (1 strong, 1 medium validity signals). 5 known CVEs in dependencies (0 critical, 3 high severity) Package registry verified. Imported from the Official MCP Registry.
5 files analyzed · 6 issues found
Security scores are indicators to help you make informed decisions, not guarantees. Always review permissions before connecting any MCP server.
This plugin requests these system permissions. Most are normal for its category.
Set these up before or after installing:
Environment variable: ANTHROPIC_API_KEY
Environment variable: CLAUDE_CODE_OAUTH_TOKEN
Environment variable: OPENAI_API_KEY
Environment variable: GOOGLE_API_KEY
Add this to your MCP configuration file:
{
"mcpServers": {
"mcp-server": {
"args": [
"-y",
"@heshamfsalama/mcp-tool-factory"
],
"command": "npx"
}
}
}From the project's GitHub README.
Generate production-ready MCP (Model Context Protocol) servers from natural language descriptions, OpenAPI specs, database schemas, GraphQL schemas, or ontologies.
The Model Context Protocol (MCP) is an open standard that enables AI assistants to securely connect with external data sources and tools. MCP servers expose tools that can be used by:
MCP Tool Factory lets you generate complete, production-ready MCP servers in seconds.
| Feature | Description |
|---|---|
| Natural Language | Describe your tools in plain English |
| OpenAPI Import | Convert any REST API spec to MCP tools |
| Database CRUD | Generate tools from SQLite or PostgreSQL schemas |
| GraphQL Import | Convert GraphQL schemas to MCP tools (queries to reads, mutations to writes) |
| Ontology Import | Generate from RDF/OWL, JSON-LD, or YAML ontologies |
| Resources & Prompts | Full support for all three MCP primitives: Tools, Resources, and Prompts |
| 10 LLM Providers | Anthropic, OpenAI, Google, Mistral, DeepSeek, Groq, xAI, Azure, Cohere + Claude Code via Vercel AI SDK |
| Cost Tracking | Per-call cost calculation, budget limits, provider cost comparison |
| Parallel Generation | Tool implementations generated concurrently for faster output |
| LLM Response Caching | Deduplicates identical LLM calls with configurable TTL |
| Streamable HTTP | Generated servers use the modern Streamable HTTP transport |
| Web Search | Auto-fetch API documentation for better generation |
| Production Ready | Logging, metrics, rate limiting, retries built-in |
| Type Safe | Full TypeScript with strict mode |
| MCP Registry | Generates server.json for registry publishing |
| Is an MCP Server | Use it directly with Claude to generate servers on-the-fly |
MCP Tool Factory is itself an MCP server! Add it to Claude Desktop, Claude Code, Cursor, or VS Code to generate MCP servers through conversation.
Claude Code auto-injects CLAUDE_CODE_OAUTH_TOKEN — no env vars needed:
claude mcp add mcp-tool-factory -- node /path/to/mcp-tool-factory-ts/bin/mcp-server.js
Set one API key and go. The factory auto-detects the provider:
Claude Desktop / Cursor / VS Code — add to your MCP config (claude_desktop_config.json, .cursor/mcp.json, or .vscode/mcp.json):
{
"mcpServers": {
"mcp-tool-factory": {
"command": "node",
"args": ["/path/to/mcp-tool-factory-ts/bin/mcp-server.js"],
"env": {
"ANTHROPIC_API_KEY": "your-key-here"
}
}
}
}
Any of these API keys will work: ANTHROPIC_API_KEY, OPENAI_API_KEY, GOOGLE_API_KEY, MISTRAL_API_KEY, DEEPSEEK_API_KEY, GROQ_API_KEY, XAI_API_KEY, AZURE_OPENAI_API_KEY, COHERE_API_KEY.
Use MCP_FACTORY_PROVIDER, MCP_FACTORY_MODEL, and MCP_FACTORY_BUDGET to override auto-detection:
{
"mcpServers": {
"mcp-tool-factory": {
"command": "node",
"args": ["/path/to/mcp-tool-factory-ts/bin/mcp-server.js"],
"env": {
"OPENAI_API_KEY": "your-key-here",
"MCP_FACTORY_PROVIDER": "openai",
"MCP_FACTORY_MODEL": "gpt-5.2",
"MCP_FACTORY_BUDGET": "0.50"
}
}
}
}
| Env Var | Purpose | Example |
|---|---|---|
MCP_FACTORY_PROVIDER | Override auto-detected provider | openai, groq, deepseek |
MCP_FACTORY_MODEL | Override default model | gpt-5.2, deepseek-chat |
MCP_FACTORY_BUDGET | Per-generation budget limit in USD | 0.50 |
Claude Code CLI with full control:
claude mcp add mcp-tool-factory \
-e DEEPSEEK_API_KEY=your-key \
-e MCP_FACTORY_PROVIDER=deepseek \
-e MCP_FACTORY_MODEL=deepseek-chat \
-e MCP_FACTORY_BUDGET=0.25 \
-- node /path/to/mcp-tool-factory-ts/bin/mcp-server.js
| Tool | Description |
|---|---|
generate_mcp_server | Generate from natural language description |
generate_from_openapi | Generate from OpenAPI specification |
generate_from_database | Generate from database schema |
generate_from_graphql | Generate from GraphQL schema |
generate_from_ontology | Generate from RDF/OWL, JSON-LD, or YAML ontology |
validate_typescript | Validate TypeScript code |
list_providers | List available LLM providers |
get_factory_info | Get factory capabilities |
You: Create an MCP server for the GitHub API with tools to list repos, create issues, and manage pull requests
Claude: Uses
generate_mcp_servertoolI've generated a complete MCP server with the following tools:
list_repositories- List user repositoriescreate_issue- Create a new issuelist_pull_requests- List PRs for a repomerge_pull_request- Merge a PRLet me write these files to your project...
# Global installation
npm install -g @heshamfsalama/mcp-tool-factory
# Or use npx
npx @heshamfsalama/mcp-tool-factory generate "Create tools for managing a todo list"
At least one provider API key is required:
# Anthropic Claude (recommended)
export ANTHROPIC_API_KEY=your-key-here
# Or Claude Code OAuth
export CLAUDE_CODE_OAUTH_TOKEN=your-token-here
# Or any other supported provider
export OPENAI_API_KEY=your-key-here
export GOOGLE_API_KEY=your-key-here
export MISTRAL_API_KEY=your-key-here
export DEEPSEEK_API_KEY=your-key-here
export GROQ_API_KEY=your-key-here
export XAI_API_KEY=your-key-here
export AZURE_OPENAI_API_KEY=your-key-here
export COHERE_API_KEY=your-key-here
# From natural language
mcp-factory generate "Create tools for fetching weather data by city and converting temperatures"
# From OpenAPI spec
mcp-factory from-openapi ./api-spec.yaml
# From database
mcp-factory from-database ./data.db
# From GraphQL schema
mcp-factory from-graphql ./schema.graphql
# From ontology
mcp-factory from-ontology ./ontology.owl --format rdf
mcp-factory generate "Create tools for managing a todo list with priorities" \
--name todo-server \
--output ./servers/todo \
--web-search \
--logging \
--metrics
# From local file
mcp-factory from-openapi ./openapi.yaml --name my-api-server
# With custom base URL
mcp-factory from-openapi ./spec.json --base-url https://api.example.com
# SQLite
mcp-factory from-database ./myapp.db --tables users,posts,comments
# PostgreSQL
mcp-factory from-database "postgresql://user:pass@localhost/mydb" --type postgresql
# From a GraphQL SDL file
mcp-factory from-graphql ./schema.graphql --name my-graphql-server
# From a URL endpoint
mcp-factory from-graphql https://api.example.com/graphql --name my-api-server
GraphQL queries are mapped to read-only MCP tools, and mutations are mapped to write tools. GraphQL types are automatically converted to Zod validation schemas.
# From RDF/OWL (.owl, .rdf, .ttl)
mcp-factory from-ontology ./ontology.owl --format rdf --name knowledge-server
# From JSON-LD (.jsonld)
mcp-factory from-ontology ./schema.jsonld --format jsonld --name linked-data-server
# From custom YAML ontology
mcp-factory from-ontology ./domain.yaml --format yaml --name domain-server
OWL Classes are mapped to MCP Resources, ObjectProperties become Tools, and DataProperties become tool parameters.
# Run tests
mcp-factory test ./servers/my-server
# Start server for testing
mcp-factory serve ./servers/my-server
servers/my-server/
├── src/
│ └── index.ts # MCP server with tools, resources, and prompts
├── tests/
│ └── tools.test.ts # Vitest tests (InMemoryTransport)
├── package.json # Dependencies
├── tsconfig.json # TypeScript config
├── Dockerfile # Container deployment
├── README.md # Usage documentation
├── skill.md # Claude Code skill file
├── server.json # MCP Registry manifest
├── EXECUTION_LOG.md # Generation trace (optional)
└── .github/
└── workflows/
└── ci.yml # GitHub Actions CI/CD
Generated servers export a createServer() factory function for easy testing. The server uses Streamable HTTP transport with a single /mcp POST endpoint and a /health GET endpoint. Tests use InMemoryTransport.createLinkedPair() for fast, reliable in-process testing with vitest.
| Command | Description |
|---|---|
generate <description> | Generate MCP server from natural language |
from-openapi <spec> | Generate from OpenAPI specification |
from-database <path> | Generate from database schema |
from-graphql <schema> | Generate from GraphQL schema |
from-ontology <file> | Generate from RDF/OWL, JSON-LD, or YAML ontology |
test <server-path> | Run tests for generated server |
serve <server-path> | Start server for testing |
info | Display factory information |
mcp-factory generate "..." \
--output, -o <path> # Output directory (default: ./servers)
--name, -n <name> # Server name
--description, -d <desc> # Package description
--github-username, -g <user> # GitHub username for MCP Registry
--version, -v <ver> # Server version (default: 1.0.0)
--provider, -p <provider> # LLM provider (anthropic, openai, google, mistral, deepseek, groq, xai, azure, cohere, claude_code)
--model, -m <model> # Specific model to use
--web-search, -w # Search web for API documentation
--auth <vars...> # Environment variables for auth
--health-check # Include health check endpoint (default: true)
--logging # Enable structured logging (default: true)
--metrics # Enable Prometheus metrics
--rate-limit <n> # Rate limiting (requests per minute)
--retries # Enable retry logic (default: true)
--budget <amount> # Maximum spend in USD (aborts if exceeded)
--compare-costs # Show cost comparison across providers before generating
| Variable | Description | Required |
|---|---|---|
ANTHROPIC_API_KEY | Anthropic Claude API key | At least one |
CLAUDE_CODE_OAUTH_TOKEN | Claude Code OAuth token | provider key |
OPENAI_API_KEY | OpenAI API key | is required |
GOOGLE_API_KEY | Google Gemini API key | for generation |
MISTRAL_API_KEY | Mistral AI API key | |
DEEPSEEK_API_KEY | DeepSeek API key | |
GROQ_API_KEY | Groq API key | |
XAI_API_KEY | xAI Grok API key | |
AZURE_OPENAI_API_KEY | Azure OpenAI API key | |
COHERE_API_KEY | Cohere API key |
All providers use the Vercel AI SDK via a unified UnifiedLLMProvider class with lazy dynamic imports — only the @ai-sdk/* package for your chosen provider is loaded at runtime.
| Provider | Models | Best For |
|---|---|---|
| Anthropic | claude-opus-4-6, claude-sonnet-4-5, claude-haiku-4-5 | Highest quality |
| OpenAI | gpt-5.2, gpt-5.2-codex, o3, o4-mini | Fast generation |
| gemini-3-pro, gemini-3-flash, gemini-2.5-pro | Cost effective | |
| Mistral | mistral-large, codestral, magistral | European AI, code |
| DeepSeek | deepseek-chat, deepseek-reasoner | Ultra low cost |
| Groq | llama-3.3-70b, llama-4-maverick | Ultra-fast inference |
| xAI | grok-4, grok-3, grok-code-fast | Reasoning |
| Azure | gpt-4o (Azure-hosted) | Enterprise compliance |
| Cohere | command-a, command-r+ | RAG, enterprise search |
| Claude Code | claude-sonnet-4-5 (OAuth) | Claude Code users |
import { ToolFactoryAgent, writeServerToDirectory, formatCost } from '@heshamfsalama/mcp-tool-factory';
// Create agent (auto-detects provider from env vars)
const agent = new ToolFactoryAgent();
// Generate from description
const server = await agent.generateFromDescription(
'Create tools for managing a todo list with priorities',
{
serverName: 'todo-server',
webSearch: true,
parallel: true, // Enable parallel generation (default)
maxConcurrency: 5, // Max concurrent LLM calls (default)
budget: 1.00, // Optional: abort if cost exceeds $1.00
productionConfig: {
enableLogging: true,
enableMetrics: true,
},
}
);
// Cost tracking — see how much the generation cost
if (server.executionLog) {
console.log(`Cost: ${formatCost(server.executionLog.totalCost)}`);
}
// Write to directory
await writeServerToDirectory(server, './servers/todo');
import { ToolFactoryAgent, writeServerToDirectory } from '@heshamfsalama/mcp-tool-factory';
import { readFileSync } from 'fs';
import yaml from 'js-yaml';
const spec = yaml.load(readFileSync('./openapi.yaml', 'utf-8'));
const agent = new ToolFactoryAgent({ requireLlm: false });
const server = await agent.generateFromOpenAPI(spec, {
serverName: 'my-api-server',
baseUrl: 'https://api.example.com',
});
await writeServerToDirectory(server, './servers/api');
import { ToolFactoryAgent, writeServerToDirectory } from '@heshamfsalama/mcp-tool-factory';
const agent = new ToolFactoryAgent({ requireLlm: false });
// SQLite (auto-detected from file path)
const server = await agent.generateFromDatabase('./data/app.db', {
serverName: 'app-database-server',
tables: ['users', 'posts', 'comments'],
});
// PostgreSQL (auto-detected from connection string)
const pgServer = await agent.generateFromDatabase(
'postgresql://user:pass@localhost/mydb',
{ serverName: 'postgres-server' }
);
await writeServerToDirectory(server, './servers/app-db');
import { ToolFactoryAgent, writeServerToDirectory } from '@heshamfsalama/mcp-tool-factory';
import { readFileSync } from 'fs';
const schema = readFileSync('./schema.graphql', 'utf-8');
const agent = new ToolFactoryAgent({ requireLlm: false });
const server = await agent.generateFromGraphQL(schema, {
serverName: 'my-graphql-server',
});
await writeServerToDirectory(server, './servers/graphql');
import { ToolFactoryAgent, writeServerToDirectory } from '@heshamfsalama/mcp-tool-factory';
import { readFileSync } from 'fs';
const ontologyData = readFileSync('./ontology.owl', 'utf-8');
const agent = new ToolFactoryAgent({ requireLlm: false });
const server = await agent.generateFromOntology(ontologyData, {
serverName: 'knowledge-server',
format: 'rdf',
});
await writeServerToDirectory(server, './servers/knowledge');
import { validateTypeScriptCode, validateGeneratedServer } from '@heshamfsalama/mcp-tool-factory';
// Validate TypeScript syntax
const result = await validateTypeScriptCode(code);
// { valid: false, errors: [{ line: 4, column: 1, message: "'}' expected." }] }
// Validate complete server
const serverResult = await validateGeneratedServer(serverCode);
// { valid: true, errors: [], summary: 'Generated server code is syntactically valid' }
Add to your MCP settings (claude_desktop_config.json):
{
"mcpServers": {
"my-server": {
"command": "npx",
"args": ["tsx", "./servers/my-server/src/index.ts"]
}
}
}
from agents import Agent
from agents.mcp import MCPServerStdio
async with MCPServerStdio(
command="npx",
args=["tsx", "./servers/my-server/src/index.ts"]
) as mcp:
agent = Agent(
name="My Agent",
tools=mcp.list_tools()
)
from google.adk.tools.mcp_tool import MCPToolset
tools = MCPToolset(
connection_params=StdioServerParameters(
command="npx",
args=["tsx", "./servers/my-server/src/index.ts"]
)
)
from langchain_mcp_adapters.client import MCPClient
client = MCPClient(
command="npx",
args=["tsx", "./servers/my-server/src/index.ts"]
)
tools = client.get_tools()
mcp-factory generate "..." --logging
Generates servers with pino structured JSON logging:
const logger = pino({ level: 'info' });
logger.info({ tool: 'get_weather', params }, 'Tool called');
mcp-factory generate "..." --metrics
Generates servers with prom-client metrics:
mcp_tool_calls_total - Counter of tool invocationsmcp_tool_duration_seconds - Histogram of execution timesmcp-factory generate "..." --rate-limit 100
Configurable rate limiting per client with sliding window.
mcp-factory generate "..." --retries
Exponential backoff retry for transient failures.
Generated servers use structured error codes for consistent error handling:
INVALID_INPUT - Malformed or invalid tool parametersNOT_FOUND - Requested resource does not existAUTH_ERROR - Authentication or authorization failureINTERNAL_ERROR - Unexpected server errorThe /health endpoint returns detailed server status:
{
"status": "ok",
"version": "1.0.0",
"uptime": 3600,
"memory": { "rss": 52428800, "heapUsed": 20971520 },
"transport": "streamable-http"
}
Publish your generated servers to the MCP Registry for discoverability.
mcp-factory generate "Create weather tools" \
--name weather-server \
--github-username your-github-username \
--description "Weather tools for Claude" \
--version 1.0.0
This generates registry-compliant files:
package.json:
{
"name": "@your-github-username/weather-server",
"mcpName": "io.github.your-github-username/weather-server"
}
server.json:
{
"$schema": "https://static.modelcontextprotocol.io/schemas/2025-12-11/server.schema.json",
"name": "io.github.your-github-username/weather-server",
"packages": [{
"registryType": "npm",
"identifier": "@your-github-username/weather-server",
"transport": { "type": "stdio" }
}],
"tools": [...]
}
# 1. Build and publish to npm
cd ./servers/weather-server
npm install && npm run build
npm publish --access public
# 2. Install mcp-publisher
brew install modelcontextprotocol/tap/mcp-publisher
# 3. Authenticate
mcp-publisher login github
# 4. Publish to registry
mcp-publisher publish
See Publishing Guide for detailed instructions.
┌───────────────────────────────────────────────────────────────────────┐
│ MCP Tool Factory │
├───────────────────────────────────────────────────────────────────────┤
│ Input Sources │
│ ┌───────────┐ ┌───────────┐ ┌───────────┐ ┌───────────┐ ┌─────────┐│
│ │ Natural │ │ OpenAPI │ │ Database │ │ GraphQL │ │Ontology ││
│ │ Language │ │ Spec │ │ Schema │ │ Schema │ │RDF/YAML ││
│ └─────┬─────┘ └─────┬─────┘ └─────┬─────┘ └─────┬─────┘ └────┬────┘│
│ └──────────┬───┴─────────────┴─────────────┴────────────┘ │
│ ▼ │
│ ┌────────────────────────────────────────────────────────────────┐ │
│ │ ToolFactoryAgent │ │
│ │ ┌─────────────────────────────────────────────────────────┐ │ │
│ │ │ UnifiedLLMProvider (Vercel AI SDK) │ │ │
│ │ │ Anthropic │ OpenAI │ Google │ Mistral │ DeepSeek │ │ │
│ │ │ Groq │ xAI │ Azure │ Cohere + Claude Code OAuth │ │ │
│ │ └─────────────────────────────────────────────────────────┘ │ │
│ │ ┌──────────────┐ ┌──────────────┐ ┌───────────────────────┐ │ │
│ │ │ LLM Cache │ │ Cost │ │ Parallel Generation │ │ │
│ │ │ (TTL-based) │ │ Tracking │ │ (max concurrency: 5) │ │ │
│ │ └──────────────┘ └──────────────┘ └───────────────────────┘ │ │
│ └────────────────────────────────────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌────────────────────────────────────────────────────────────────┐ │
│ │ Generators │ │
│ │ ServerGenerator │ DocsGenerator │ TestsGenerator │ │
│ └────────────────────────────────────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌────────────────────────────────────────────────────────────────┐ │
│ │ GeneratedServer │ │
│ │ Tools │ Resources │ Prompts │ Tests │ Docs │ Dockerfile │ │
│ └────────────────────────────────────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌────────────────────────────────────────────────────────────────┐ │
│ │ Streamable HTTP Transport │ │
│ │ POST /mcp │ GET /health │ │
│ └────────────────────────────────────────────────────────────────┘ │
└───────────────────────────────────────────────────────────────────────┘
# Clone the repository
git clone https://github.com/HeshamFS/mcp-tool-factory-ts.git
cd mcp-tool-factory-ts
# Install dependencies
pnpm install
# Build
pnpm run build
# Run tests
pnpm test
# Type check
pnpm run typecheck
# Lint
pnpm run lint
mcp-tool-factory-ts/
├── src/
│ ├── agent/ # Main ToolFactoryAgent
│ ├── auth/ # OAuth2 providers
│ ├── cache/ # LLM response caching with configurable TTL
│ ├── cli/ # Command-line interface
│ ├── config/ # Configuration management
│ ├── database/ # Database introspection (SQLite, PostgreSQL)
│ ├── execution-logger/ # Execution logging
│ ├── generators/ # Code generators (server, docs, tests)
│ ├── graphql/ # GraphQL SDL parsing and server generation
│ ├── middleware/ # Validation middleware
│ ├── models/ # Data models
│ ├── observability/ # Telemetry and tracing
│ ├── ontology/ # Ontology parsing (RDF/OWL, JSON-LD, YAML)
│ ├── openapi/ # OpenAPI spec parsing
│ ├── production/ # Production code generation
│ ├── prompts/ # LLM prompt templates
│ ├── providers/ # LLM providers (10 providers via Vercel AI SDK + Claude Code)
│ ├── security/ # Security scanning
│ ├── server/ # MCP server mode (factory-as-a-server)
│ ├── templates/ # Handlebars templates for generated files
│ ├── validation/ # Code validation and Zod schemas
│ └── web-search/ # Web search integration
├── docs/ # Documentation
├── tests/ # Test files
└── dist/ # Built output
API Key Not Found
# Check your environment
echo $ANTHROPIC_API_KEY
# Set it
export ANTHROPIC_API_KEY=your-key-here
Generated Server Won't Start
# Install dependencies first
cd ./servers/my-server
npm install
npx tsx src/index.ts
TypeScript Errors
# Validate generated code
import { validateGeneratedServer } from '@heshamfsalama/mcp-tool-factory';
const result = await validateGeneratedServer(code);
console.log(result.errors);
See Troubleshooting Guide for more solutions.
UnifiedLLMProvider class with lazy dynamic imports. Removed ~473 LOC of provider-specific implementations. Only the @ai-sdk/* package for your chosen provider is loaded at runtime.--budget <amount>) - Set a maximum spend in USD. Generation aborts gracefully with BudgetExceededError if cumulative cost exceeds the budget.--compare-costs) - Before generation, estimates cost across all available providers and shows a sorted comparison table. No extra API calls needed — uses the static pricing table.StreamableHTTPServerTransport with native http module instead of Express/SSE (deprecated June 2025). Single /mcp POST endpoint with /health GET endpoint.^1.0.0 to ^1.26.0from-graphql CLI command and generate_from_graphql MCP tool. Queries map to read tools, mutations map to write tools, and GraphQL types are converted to Zod schemas.from-ontology CLI command and generate_from_ontology MCP tool. Supports RDF/OWL, JSON-LD, and custom YAML formats. OWL Classes map to Resources, ObjectProperties to Tools, DataProperties to tool parameters.skipCache option.parallel: true, maxConcurrency: 5). Significant speed improvement for multi-tool servers.InMemoryTransport.createLinkedPair() instead of subprocess spawning. Servers export createServer() factory function for testability.INVALID_INPUT, NOT_FOUND, AUTH_ERROR, INTERNAL_ERROR.MIT
Be the first to review this server!
by Modelcontextprotocol · Developer Tools
Read, search, and manipulate Git repositories programmatically
by Toleno · Developer Tools
Toleno Network MCP Server — Manage your Toleno mining account with Claude AI using natural language.
by mcp-marketplace · Developer Tools
Create, build, and publish Python MCP servers to PyPI — conversationally.
by Microsoft · Content & Media
Convert files (PDF, Word, Excel, images, audio) to Markdown for LLM consumption
by mcp-marketplace · Developer Tools
Scaffold, build, and publish TypeScript MCP servers to npm — conversationally
by mcp-marketplace · Finance
Free stock data and market news for any MCP-compatible AI assistant.