loading…
Search for a command to run...
loading…
An MCP server that functions as an intelligent gateway for multiple LLM backends including OpenAI, Claude, and Ollama. It supports automatic provider fallback,
An MCP server that functions as an intelligent gateway for multiple LLM backends including OpenAI, Claude, and Ollama. It supports automatic provider fallback, streaming responses via Server-Sent Events, and real-time monitoring for robust AI integration.
Multi-LLM Gateway para mcp.observabilidadebrasil.org
Um servidor MCP (Model Context Protocol) que atua como gateway inteligente para múltiplos backends LLM, com suporte a streaming, rate limiting, e monitoramento.
┌─────────────────────────────────────────────────────────────┐
│ nginx (SSL + Rate Limit) │
│ mcp.observabilidadebrasil.org │
└─────────────────────────┬───────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ MCP Server (FastAPI) │
│ Port 9200 │
├─────────────────────────────────────────────────────────────┤
│ ┌─────────┐ ┌─────────┐ ┌─────────┐ ┌─────────┐ │
│ │ OpenAI │ │ Claude │ │ Ollama │ │ Custom │ │
│ │ Provider│ │ Provider│ │ Provider│ │ Provider│ │
│ └────┬────┘ └────┬────┘ └────┬────┘ └────┬────┘ │
│ │ │ │ │ │
│ └────────────┴────────────┴────────────┘ │
│ │ │
│ ┌──────────▼──────────┐ │
│ │ LLM Router │ │
│ │ (load balance, │ │
│ │ fallback, routing) │ │
│ └─────────────────────┘ │
└─────────────────────────────────────────────────────────────┘
# Clonar repositório
git clone https://github.com/tgosoul2019/mcp.git
cd mcp
# Criar virtual environment
python3 -m venv .venv
source .venv/bin/activate
# Instalar dependências
pip install -e ".[dev]"
# Configurar variáveis de ambiente
cp .env.example .env
# Editar .env com suas API keys
# Rodar servidor
python -m mcp_server
# No servidor
cd /dados
git clone https://github.com/tgosoul2019/mcp.git
cd mcp
# Setup
./scripts/setup.sh
# Iniciar serviço
sudo systemctl start mcp-server
# Server
MCP_HOST=127.0.0.1
MCP_PORT=9200
MCP_DEBUG=false
# LLM Providers (configure apenas os que usar)
OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
OLLAMA_BASE_URL=http://localhost:11434
# Default Provider
MCP_DEFAULT_PROVIDER=openai
# Rate Limiting (aplicação)
MCP_RATE_LIMIT_REQUESTS=100
MCP_RATE_LIMIT_WINDOW=60
# Logging
MCP_LOG_LEVEL=INFO
MCP_LOG_FILE=/var/log/mcp/mcp.log
POST /v1/chat/completions
Content-Type: application/json
{
"model": "gpt-4",
"messages": [
{"role": "user", "content": "Hello!"}
],
"stream": true,
"provider": "openai" # opcional, usa default se omitido
}
GET /health
GET /health/providers
GET /metrics
O MCP tem seu próprio dashboard de monitoramento separado do KCP:
https://mcp.observabilidadebrasil.org/admin/monitor# Build
docker build -t mcp-server .
# Run
docker run -d \
--name mcp-server \
-p 9200:9200 \
-e OPENAI_API_KEY=sk-... \
mcp-server
mcp/
├── mcp_server/
│ ├── __init__.py
│ ├── __main__.py
│ ├── app.py # FastAPI app
│ ├── config.py # Configurações
│ ├── router.py # LLM Router
│ ├── providers/
│ │ ├── __init__.py
│ │ ├── base.py # Abstract Provider
│ │ ├── openai.py
│ │ ├── anthropic.py
│ │ └── ollama.py
│ ├── middleware/
│ │ ├── __init__.py
│ │ ├── rate_limit.py
│ │ └── logging.py
│ └── monitor/
│ ├── __init__.py
│ ├── collector.py # Métricas
│ └── dashboard.py # UI
├── infra/
│ ├── nginx/
│ │ └── mcp.conf
│ ├── systemd/
│ │ └── mcp-server.service
│ └── docker/
│ ├── Dockerfile
│ └── docker-compose.yml
├── scripts/
│ ├── setup.sh
│ └── deploy.sh
├── tests/
├── pyproject.toml
├── .env.example
└── README.md
MIT
Добавь это в claude_desktop_config.json и перезапусти Claude Desktop.
{
"mcpServers": {
"multi-llm-gateway-mcp": {
"command": "npx",
"args": []
}
}
}Web content fetching and conversion for efficient LLM usage.
Retrieval from AWS Knowledge Base using Bedrock Agent Runtime.
Provides auto-configuration for setting up an MCP server in Spring Boot applications.
A very streamlined mcp client that supports calling and monitoring stdio/sse/streamableHttp, and can also view request responses through the /logs page. It also