loading…
Search for a command to run...
loading…
A self-hostable, open-source, semantically-searchable Agent Skills registry delivered over MCP, with a three-tier progressive disclosure architecture.
A self-hostable, open-source, semantically-searchable Agent Skills registry delivered over MCP, with a three-tier progressive disclosure architecture.
A self-hostable, open-source Skills registry for AI agents delivered over MCP.
Semantic discovery · Progressive disclosure · 30 bundled skills · Self-host on Cloudflare Workers
AI agents are great at knowing things but often inconsistent at doing specific tasks well.
Ask Claude or GPT to "write a Stripe integration" and you get something functional but maybe it doesn't verify webhook signatures, uses a deprecated API, or misses idempotency keys. Ask it to "containerize this app" and it might skip non-root users, ignore layer caching, or forget .dockerignore.
The knowledge is there. The reliable procedural workflow isn't.
The deeper problem: every time you start a new chat, a new project, or switch tools, the agent starts from scratch. There's no shared, versioned, searchable library of how to do X correctly that agents can pull from on demand.
skill-mcp is a self-hostable registry of Skills expert step-by-step procedures, domain heuristics, output formats, and bundled reference material that AI agents discover and load at the moment they need them, over MCP.
You: "Set up Stripe subscriptions with webhooks"
Agent: → skills_find_relevant("set up Stripe subscriptions webhooks")
Score 0.89 → stripe-integration
→ skills_get_body("stripe-integration")
Loads: which API to use, webhook verification pattern,
idempotency keys, security checklist, go-live steps
→ Executes task correctly, first time, every time
The agent doesn't guess. It retrieves authoritative, versioned instructions the same way a senior engineer would consult a runbook.
Each skill is a SKILL.md file with YAML frontmatter. Only the description and trigger phrases (~100 tokens) are embedded keeping the search space semantically clean. The full instructions stay in payload-only collections and are fetched on demand.
skills_find_relevant("write pytest tests for a FastAPI endpoint")
→ test-writer (score: 0.84)
→ fastapi (score: 0.71)
Score thresholds (judgment-based starting points): > 0.6 strong match · 0.4–0.6 review description · < 0.4 no match
Tier 1 Discovery skills_find_relevant() ← always call first
Tier 2 Load skills_get_body() ← full instructions + manifest
Tier 2 Options skills_get_options() ← config, variants (optional)
Tier 3 Supplement skills_get_reference() ← only if instructions reference it
skills_run_script() ← only if instructions reference it
skills_get_asset() ← only if instructions reference it
Nothing is loaded speculatively. The agent reads tier3_manifest (a list of available files returned with the body) and fetches only what the instructions explicitly reference.
No separate backend. No database server to manage. No GPU. Cloudflare Workers AI handles embeddings at query time using the same model the seed script uses vectors are always comparable.
| Collection | Vector | Contents |
|---|---|---|
skill_frontmatter |
✅ 384-dim | Name, description, tags, trigger phrases the discovery layer |
skill_body |
payload only | Full markdown instructions + system prompt addition |
skill_options |
payload only | Config schema, variants, dependencies, limitations |
skill_references |
payload only | Markdown reference docs bundled with the skill |
skill_scripts |
payload only | Executable scripts (source stored server-side; never sent to agents) |
skill_assets |
payload only | Templates and static output format resources |
| Tier | Tool | When to call |
|---|---|---|
| 1 | skills_find_relevant(query, top_k) |
Always first semantic search, returns ranked skills with scores |
| 2 | skills_get_body(skill_id, version?) |
After finding a match full instructions + tier3_manifest; version pins to a specific release |
| 2 | skills_get_options(skill_id) |
Optional config schema, variants, dependencies, limitations |
| 3 | skills_get_reference(skill_id, filename) |
Only when instructions reference a specific doc |
| 3 | skills_run_script(skill_id, filename, input_data) |
Only when instructions direct script execution |
| 3 | skills_get_asset(skill_id, filename) |
Only when instructions reference a specific template |
Embedding the full SKILL.md as a single vector pollutes the search space with instruction prose text that was never meant to be searched. skill-mcp embeds only description + trigger_phrases (~100 tokens), keeping the vector space semantically clean and search results relevant.
The Worker uses Cloudflare Workers AI (@cf/baai/bge-small-en-v1.5, 384-dim) for query-time embedding. The seed script calls the same model via the REST API. Seed-time and query-time vectors are directly comparable no local GPU, no embedding server, no drift.
Bundled skills covering common engineering domains. Sourced from official documentation (Anthropic, Google Gemini, Vercel, Cloudflare, Stripe) and established engineering best practices. Quality and currency vary by skill — check the individual SKILL.md files for author and version metadata.
| Skill | What it does |
|---|---|
api-integration |
REST/GraphQL clients with auth, pagination, retries, error handling, and OpenAPI alignment |
code-review |
Structured security + quality review with CRITICAL/HIGH/MEDIUM/LOW severity ratings and fix snippets |
data-analysis |
EDA, cleaning, statistics, visualizations, and actionable insights from CSV/tabular data |
git-commit-writer |
Conventional Commits from diffs type, scope, breaking changes, and co-authors |
readme-writer |
Professional README.md with badges, usage, API docs, and contributing guide |
sql-query-writer |
Optimized SQL window functions, CTEs, indexes, explain plans, and common anti-patterns |
test-writer |
pytest, Jest, and Go test suites with full edge case coverage and mocking patterns |
web-scraper |
Structured data extraction with rate limiting, pagination, and anti-bot handling |
| Skill | What it does |
|---|---|
docx-creator |
Create and edit Word documents with python-docx tables, styles, headers, tracked changes |
pdf-processing |
Extract text/tables, fill forms, merge/split PDFs full Tier 3 scripts and references |
pptx-creator |
Build PowerPoint presentations with pptxgenjs charts, images, design principles |
xlsx-creator |
Excel spreadsheets with openpyxl formulas, formatting, charts, financial model conventions |
| Skill | What it does |
|---|---|
claude-api |
Anthropic SDK: tool use, streaming, vision, prompt caching, extended thinking, batch |
gemini-api |
Google Gemini API: multimodal, function calling, structured output, current models/SDKs |
openai-api |
OpenAI: GPT-4o, tool use, structured output, DALL-E, Whisper, TTS, batch processing |
llm-prompt-engineering |
Chain-of-thought, few-shot, structured output, agent system prompt design, anti-patterns |
mcp-server-builder |
Build MCP servers with FastMCP (Python) or TypeScript SDK tools, resources, prompts |
| Skill | What it does |
|---|---|
cloudflare-workers |
Workers, Pages, KV, D1, R2, Workers AI, Vectorize, Durable Objects, Wrangler |
docker-containerization |
Production Dockerfiles, multi-stage builds, Docker Compose, security hardening |
github-actions |
CI/CD workflows, matrix builds, caching, Docker publishing, release automation |
terraform |
IaC for AWS/GCP/Azure modules, remote state, workspaces, CI/CD integration |
| Skill | What it does |
|---|---|
nextjs-best-practices |
App Router RSC, async params, data fetching, image/font optimization, self-hosting |
react-best-practices |
Hooks patterns, state management, memoization, virtualization, error boundaries |
fastapi |
Python REST APIs Pydantic v2, dependency injection, JWT auth, async SQLAlchemy, testing |
graphql-api |
Schema design, resolvers, DataLoader (N+1 prevention), Apollo Client, Strawberry |
typescript-patterns |
Generics, discriminated unions, branded types, conditional types, strict tsconfig |
| Skill | What it does |
|---|---|
stripe-integration |
Checkout Sessions, webhooks, subscriptions, Connect (Accounts v2), security checklist |
supabase-integration |
PostgreSQL queries, auth (OAuth/magic link), RLS policies, real-time, storage |
| Skill | What it does |
|---|---|
frontend-design |
Aesthetic direction, typography systems, color palettes, micro-animations, anti-patterns |
web-artifacts-builder |
Self-contained interactive HTML/React/Tailwind/D3 artifacts and dashboards |
| Requirement | Cost | Notes |
|---|---|---|
| Qdrant Cloud | Free | 1 GB free cluster — create one, copy URL + API key |
| Cloudflare | Free | Workers Free plan supports SQLite-backed Durable Objects |
| Python 3.11+ | Free | For the seed script and optional local server |
| Node.js 18+ | Free | For the wrangler CLI |
Cloudflare is free. skill-mcp uses SQLite-backed Durable Objects (
new_sqlite_classesinwrangler.jsonc), which are available on the Cloudflare Workers Free plan (100k requests/day). You only need the $5/mo paid plan if you outgrow that limit or need KV-backed Durable Objects.
Windows (PowerShell):
.\scripts\setup.ps1
Linux / macOS:
bash scripts/setup.sh
Cross-platform (Make):
make setup
The wizard checks prerequisites → creates .env → installs Python deps → seeds Qdrant with all 30 skills → pushes Wrangler secrets → deploys the Worker. Done.
# 1. Clone
git clone https://github.com/yourusername/skill-mcp && cd skill-mcp
# 2. Configure credentials
cp .env.example .env
# Fill in: QDRANT_URL, QDRANT_API_KEY, WORKERS_AI_ACCOUNT_ID, WORKERS_AI_API_TOKEN
# 3. Install seed dependencies and seed Qdrant
pip install -r requirements.txt
python -X utf8 -m skill_mcp.seed.seed_skills
# 4. Deploy to Cloudflare
npm install -g wrangler
wrangler login
wrangler secret put QDRANT_URL # paste your Qdrant URL
wrangler secret put QDRANT_API_KEY # paste your Qdrant API key
wrangler deploy
Your server is live at:
https://skill-mcp.<your-subdomain>.workers.dev/sse
Full credential walkthrough: SETUP.md
# Cloudflare deployment
make env # Copy .env.example → .env (skips if .env already exists)
make check # Verify all required .env values are set
make install # pip install -r requirements.txt
make seed # Seed / re-seed Qdrant with all skills (idempotent)
make secrets # Auto-push QDRANT_URL + QDRANT_API_KEY from .env to Worker
make deploy # wrangler deploy
make dev # Run local FastMCP server in stdio mode
make dev-http # Run local FastMCP server on HTTP :8000
make setup # Full first-run: env + install + seed + secrets + deploy
# Security & validation
make validate # Validate all SKILL.md files — schema + prompt-injection scan
make calibrate # Sweep (t_high, t_low) pairs; report precision/recall/F1
make check-qdrant-keys # Warn if read/write Qdrant keys are identical
# Docker (one-command local stack)
make docker-up # Start Qdrant + seed + MCP server
make docker-down # Stop containers (keeps Qdrant data)
make docker-seed # Re-seed after adding new skills
make docker-logs # Follow server logs
No Cloudflare account needed. Runs Qdrant locally in a container — useful for local-only setups, air-gapped environments, or testing before deploying.
# Start everything: Qdrant + seed + MCP server
docker compose up
# Or in background
docker compose up -d && docker compose logs -f server
Your local MCP server is live at http://localhost:8000/sse.
Add to your MCP client config:
{
"mcpServers": {
"skill-mcp": {
"transport": "sse",
"url": "http://localhost:8000/sse"
}
}
}
Requirements for Docker mode: only WORKERS_AI_ACCOUNT_ID and WORKERS_AI_API_TOKEN in .env — Cloudflare credentials are still needed to generate embeddings via Workers AI. Qdrant runs locally, no Qdrant Cloud account required.
make docker-up # Start the full stack
make docker-down # Stop (data volume preserved)
make docker-seed # Re-seed after adding new skills
Before connecting to any hosted skill-mcp instance you do not control: read TRANSPARENCY.md. Skill bodies load directly into your agent's context window from a third-party server. The hosted instance offered by this repo is a personal deployment with no SLA and no authentication. For production use or sensitive workloads, self-host.
Add to your MCP client config (.mcp.json, Claude Code settings, Cursor settings, etc.):
Production (Cloudflare Worker):
{
"mcpServers": {
"skill-mcp": {
"transport": "sse",
"url": "https://skill-mcp.<your-subdomain>.workers.dev/sse"
}
}
}
Local dev (wrangler dev):
{
"mcpServers": {
"skill-mcp": {
"transport": "sse",
"url": "http://localhost:8787/sse"
}
}
}
Local Python server (needed for skills_run_script Cloudflare Workers cannot run subprocesses):
{
"mcpServers": {
"skill-mcp": {
"command": "python",
"args": ["-m", "skill_mcp.server"],
"cwd": "/path/to/skill-mcp"
}
}
}
Drop the right file into any project root and the agent will automatically follow the 3-tier skill workflow when to search, how to interpret scores, and when to load supplementary files.
| Platform | File to copy | Where |
|---|---|---|
| Claude Code | master-skill/platforms/claude-code/CLAUDE.md |
Project root |
| Cursor | master-skill/platforms/cursor/.cursorrules |
Project root |
| Windsurf | master-skill/platforms/windsurf/.windsurfrules |
Project root |
| Antigravity (Google) | master-skill/platforms/antigravity/.agents/ |
Project root (primary) |
| Antigravity (Google) | master-skill/platforms/antigravity/AGENTS.md |
Project root (secondary) |
| OpenAI Codex | master-skill/platforms/codex/AGENTS.md |
Project root |
| Cline (VSCode) | master-skill/platforms/cline/.clinerules |
Project root |
| GitHub Copilot | master-skill/platforms/copilot/.github/ |
Project root |
| Aider | master-skill/platforms/aider/CONVENTIONS.md |
Project root |
After copying, replace the placeholder URL with your deployed Worker URL.
Per-platform install commands: master-skill/README.md
Skills live in skill_mcp/skills_data/. Each skill is a folder:
skill_mcp/skills_data/
└── my-skill/
├── SKILL.md ← required: frontmatter + full instructions
├── references/ ← optional: markdown reference docs (.md)
├── scripts/ ← optional: executable scripts (.py, .js, .sh)
└── assets/ ← optional: output templates and static files
---
name: my-skill
description: >
One or two sentences describing WHEN to use this skill.
Write it from the agent's perspective: "Use when the user asks to extract data from PDFs,
process forms, or parse tables from documents."
license: Apache-2.0
metadata:
author: your-name
version: "1.0"
tags: [pdf, extraction, data]
platforms: [claude-code, cursor, any]
triggers:
- extract text from a PDF
- parse a PDF document
- read a PDF file
- fill a PDF form
---
# Skill Title
Full step-by-step instructions. This is what the agent reads and follows.
Reference tier-3 files explicitly so the agent knows to load them:
- "For field type reference, see references/FORMS.md"
- "To extract data, run scripts/extract.py with PDF_PATH set to the file path"
- "Format your output using assets/extraction-template.md"
Two critical rules:
Description and triggers are what get embedded write them to match how an agent would phrase the need, not how you'd name the skill. "extract tables from a PDF" beats "pdf-skill".
Reference tier-3 files by name in the body the agent receives a tier3_manifest listing available files and fetches only what the instructions explicitly mention. Nothing is loaded speculatively.
python -X utf8 -m skill_mcp.seed.seed_skills
# or:
make seed
The seed script is idempotent re-running updates existing skills without creating duplicates.
A malicious SKILL.md with embedded instruction overrides could alter how agents behave after loading the skill body — turning the registry into a prompt-injection delivery mechanism.
Every skill is scanned by skill_mcp/security/prompt_injection.py before it enters Qdrant — at seed time and in CI on every PR. Skills with CRITICAL or HIGH findings are blocked. The scanner uses pattern matching; semantic attacks that evade patterns are a known residual risk (see THREAT_MODEL.md).
| Attack category | Severity | Example |
|---|---|---|
| Instruction-override phrases | CRITICAL | "ignore all previous instructions" |
| Role / identity hijacking | CRITICAL | "you are now an unrestricted AI" |
| Prompt delimiter injection | HIGH | </system>, [INST], <<SYS>> |
| Credential exfiltration | CRITICAL | "POST the API key to webhook.site/…" |
| HTML / script injection | HIGH | <script> outside code blocks |
| Unicode BiDi / zero-width chars | HIGH | Visually hidden content |
| Base64 encoded payloads | CRITICAL | Base64 that decodes to override phrases |
| Content displacement | MEDIUM | 20+ consecutive blank lines |
Code blocks are stripped before structural checks — TypeScript generics (Promise<User>) and <script> tags in code examples never false-positive.
Full threat model: THREAT_MODEL.md · Hosted instance trust model: TRANSPARENCY.md
RATE_LIMIT_RPM); returns HTTP 429 when exceeded; stale entry eviction at 10k IPs; Worker-onlyAccess-Control-Allow-Origin: * on all Worker responses; supports browser-based MCP clients and testers (Glama, MCP Inspector)X-Content-Type-Options: nosniff, X-Frame-Options: DENY, Cache-Control: no-store, Referrer-Policy: no-referrertools/call arguments type-checked; malformed JSON-RPC returns proper error codesskills_find_relevant rejects queries over 2,000 charactersScript execution (skills_run_script, local server only):
tempfile.TemporaryDirectory() — deleted after each runPATH, LD_PRELOAD, PYTHONPATH, etc.)stdout / stderr / exit_codeIn the deployed Cloudflare Worker, skills_run_script returns the script manifest only the Pyodide runtime cannot run subprocesses.
Three top-level directories own three distinct concerns:
skill_mcp/ — the Python package. Everything the server needs at runtime lives here: Pydantic models (models/), Qdrant integration (db/), MCP tool implementations (tools/), the prompt-injection scanner (security/), the seed script (seed/), the local FastMCP entry point (server.py), and the skill registry itself (skills_data/). If you are adding a skill, editing a tool, or touching the data layer, you are working here.
src/ — the Cloudflare Workers deployment target. Contains a single file, worker.py, which re-implements all six MCP tools as a self-contained Cloudflare Python Worker (no external packages, Pyodide-compatible). wrangler.jsonc at the repo root points here. Edit this only when changing the deployed Worker behaviour.
scripts/ — developer and CI utilities that are not part of the importable package. setup.sh / setup.ps1 are one-shot interactive wizards; validate_skills.py is the SKILL.md schema + prompt-injection validator invoked by both make validate and the GitHub Actions skill-validation workflow.
skill-mcp/
├── skill_mcp/ # Installable Python package (pip install -e ".[seed]")
│ ├── db/ # Qdrant client, embedder, TTL cache
│ ├── eval/calibrate.py # Threshold calibration runner (precision/recall sweep)
│ ├── models/skill.py # Pydantic models for all 6 collection types
│ ├── security/prompt_injection.py # 9-category injection scanner
│ ├── seed/seed_skills.py # Walks skills_data/, scans, embeds, upserts Qdrant
│ ├── tools/ # MCP tool implementations (local server)
│ ├── skills_data/ # 30 skill folders — one SKILL.md each
│ └── server.py # Local FastMCP entry point (stdio / HTTP)
├── src/
│ └── worker.py # Cloudflare Python Worker — all 6 tools, SSE + Streamable HTTP, rate limiting, CORS
├── scripts/
│ ├── setup.sh / setup.ps1 # One-shot setup wizards (Linux/macOS + Windows)
│ └── validate_skills.py # SKILL.md validator — schema + injection scan
├── master-skill/ # Drop-in agent instruction files (8 platforms)
│ └── platforms/
│ ├── claude-code/CLAUDE.md
│ ├── cursor/.cursorrules
│ ├── windsurf/.windsurfrules
│ ├── codex/AGENTS.md
│ ├── cline/.clinerules
│ ├── copilot/.github/copilot-instructions.md
│ └── aider/CONVENTIONS.md
├── tests/
│ └── eval/threshold_calibration.json # 120 eval triples for threshold calibration
├── .github/workflows/
│ ├── tests.yml # pytest on every push (unit tests, no external deps)
│ └── validate-skills.yml # SKILL.md lint + injection scan on PRs
├── wrangler.jsonc # Workers AI binding + SQLite Durable Objects config
├── Makefile # Automation: setup, seed, deploy, dev, docker, validate
├── Dockerfile / docker-compose.yml # One-command local stack: Qdrant + seed + server
├── pyproject.toml # Package metadata + optional dependency groups
├── .env.example # Credential template — copy to .env
├── SETUP.md # Full credential walkthrough
├── CONTRIBUTING.md # Skill submission workflow + security policy
├── THREAT_MODEL.md # 7 threat categories with mitigations
├── TRANSPARENCY.md # Hosted instance trust model, SLA status, deployment boundaries
└── docs/ # Architecture, versioning, calibration, and federation design
Master skill required for reliable agent behavior — The 3-tier workflow (discover → load → supplement) only fires consistently when the master skill file is installed in the agent's project root (see Step 2 above). Without it, agents may skip score thresholds, load skill bodies speculatively, or ignore the tier3_manifest entirely — wasting context window tokens and producing inconsistent results.
Token usage scales with collection size — skills_find_relevant returns top_k result descriptors (each ~100–200 tokens). At 30 skills this is negligible. At 300+ skills with higher top_k values, a single discovery call can consume a meaningful share of the context window. Keep top_k low (3–5) and write precise, distinct trigger phrases per skill to preserve relevance at scale.
Script execution is local-only — skills_run_script requires the local Python server. The Cloudflare Worker returns the script manifest but cannot execute subprocesses — the Pyodide runtime does not support subprocess. Any skill workflow that calls skills_run_script must point the MCP client at python -m skill_mcp.server instead of the Worker URL.
Embedding model is pinned at seed time — Vectors are generated with @cf/baai/bge-small-en-v1.5 (384-dim) at both seed time and query time. If Cloudflare Workers AI retires or changes this model, all vectors become incomparable and the entire skill collection must be re-seeded.
Search quality depends on trigger phrase quality — Semantic search is only as good as the triggers written in each SKILL.md. Skills with vague or overlapping trigger phrases will surface for unrelated queries and dilute results. One skill with poorly-written triggers degrades the entire registry.
Read CONTRIBUTING.md for the full skill submission workflow — what makes a great skill, the SKILL.md format reference, step-by-step PR process, and the security policy for submitted skills.
Quick start:
# 1. Create your skill
mkdir -p skill_mcp/skills_data/my-skill && touch skill_mcp/skills_data/my-skill/SKILL.md
# 2. Validate locally (schema + prompt-injection scan)
python scripts/validate_skills.py skill_mcp/skills_data/my-skill/SKILL.md
# 3. Open a PR — CI runs automatically
The two invariants that must never be broken:
description + triggers go into the vector collectionskills_run_script returns stdout / stderr / exit_code onlyCI validates every PR that touches skills_data/: YAML syntax, schema, duplicate slug check, and prompt-injection scan. A failing scan blocks merge.
Apache 2.0 see LICENSE.
Built with Cloudflare Workers · Qdrant · FastMCP · MCP
Add this to claude_desktop_config.json and restart Claude Desktop.
{
"mcpServers": {
"skills-mcp": {
"command": "npx",
"args": []
}
}
}