loading…
Search for a command to run...
loading…
Multi-provider search broker with automatic fallback, RRF ranking, content extraction, and budget enforcement.
Multi-provider search broker with automatic fallback, RRF ranking, content extraction, and budget enforcement.
Python 3.11+ PyPI Version PyPI Downloads License: MIT CI MCP Registry Docker
Multi-provider web search broker for AI agents. 14 providers, budget-aware routing, content extraction — one API so your agent doesn't need to stitch search results together.
Features at a glance:
pip install argus-search gives you DuckDuckGo + Yahoo immediately, no accounts neededsession_id for conversational context across searches/recover-url with Wayback Machine and archive fallbacksBuilt for AI agent builders, RAG pipelines, and ops teams who need reliable search without stitching APIs together.
pip install argus-search && argus search -q "python web frameworks"
That's it. DuckDuckGo handles the search — no accounts, no keys, no containers. You get unlimited free search from your laptop right now. Add API keys whenever you want more providers, or don't.
argus extract -u "https://example.com/article" # extract clean text from any URL
Works on any machine with Python 3.11+ — laptop, Mac Mini, Raspberry Pi, cloud VM. Nothing to host.
For MCP (Claude Code, Cursor, VS Code):
pipx install argus-search[mcp] && argus mcp serve
Then add to your MCP config:
{"mcpServers": {"argus": {"command": "argus", "args": ["mcp", "serve"]}}}
Or install from the MCP Registry:
{
"mcpServers": {
"argus": {
"registryType": "pypi",
"identifier": "argus-search",
"runtimeHint": "uvx"
}
}
}
One command to install, one JSON block to connect. No server to run, no keys to configure.
Got a Raspberry Pi running Pi-hole? A Mac Mini on your desk? An old laptop? That's enough to run the full stack — SearXNG (your own private search engine) plus local JS-rendering content extraction.
docker compose up -d # SearXNG + Argus
| What you have | What you get |
|---|---|
| Any machine with Python 3.11+ | DuckDuckGo + API providers (no server) |
| Raspberry Pi 4 / old laptop (4GB+) | Everything — SearXNG, all providers, Crawl4AI, Obscura |
| Mac Mini M1+ (8GB+) | Full stack with headroom |
| Free cloud VM (1GB) | SearXNG + search providers (skip Crawl4AI) |
SearXNG takes 512MB of RAM and gives you a private Google-style search engine that nobody can rate-limit, block, or charge for. It runs alongside Pi-hole on hardware millions of people already own.
Repo development is pinned to Python 3.12. The package runtime floor remains Python 3.11, but contributors should use the uv workflow below so local verification matches CI and avoids accidentally using an older system interpreter.
uv sync --python 3.12 --extra dev --extra mcp
uv run pytest tests/ -v --tb=short
The repo includes .python-version with 3.12 so uv, pyenv, and similar tools pick the right interpreter by default. More contributor guidance lives in CONTRIBUTING.md.
| Provider | Credit type | Free capacity | Setup |
|---|---|---|---|
| DuckDuckGo | Free (scraped) | Unlimited | None |
| Yahoo | Free (scraped) | Unlimited | None — fragile, auto-skipped if broken |
| SearXNG | Free (self-hosted) | Unlimited — 70+ engines¹ | Docker |
| GitHub | Free (API) | Unlimited | None (token for higher rate limit) |
| WolframAlpha | Free (API key) | 2,000 queries/month | free key |
| Brave Search | Monthly recurring | 2,000 queries/month | dashboard |
| Tavily | Monthly recurring | 1,000 queries/month | signup |
| Exa | Monthly recurring | 1,000 queries/month | signup |
| Linkup | Monthly recurring | 1,000 queries/month | signup |
| Serper | One-time signup | 2,500 credits | signup |
| Parallel AI | One-time signup | 4,000 credits | signup |
| You.com | One-time signup | $20 credit | platform |
| Valyu | One-time signup | $10 credit | platform |
¹ SearXNG aggregates Google, Bing, Yahoo, Startpage, Ecosia, Qwant, Wikipedia, and 60+ more — all behind a single self-hosted endpoint. Run docker compose up -d on any machine with 512MB of free RAM.
² WolframAlpha returns computed answers (math, unit conversions, factual lookups), not web search results. It only activates in grounding and research modes. Queries it can't compute (general web searches) return empty — no error, no health penalty.
7,000+ free queries/month from recurring free-tier providers alone (WolframAlpha 2k + Brave 2k + Tavily 1k + Exa 1k + Linkup 1k). DuckDuckGo, Yahoo, SearXNG, and GitHub have no monthly cap. Routing priority: Tier 0 (free: SearXNG, DuckDuckGo, Yahoo, GitHub, WolframAlpha) → Tier 1 (monthly recurring: Brave, Tavily, Exa, Linkup) → Tier 3 (one-time: Serper, Parallel, You.com, Valyu, SearchAPI). Budget-exhausted providers are skipped automatically.
All endpoints prefixed with /api. OpenAPI docs at http://localhost:8000/docs.
Local loopback calls can use the API without auth. Remote HTTP callers must send ARGUS_API_KEY as either Authorization: Bearer ... or X-API-Key: .... Privileged routes under /api/admin/* require ARGUS_ADMIN_API_KEY (or fall back to ARGUS_API_KEY if no separate admin key is configured).
# Search
curl -X POST http://localhost:8000/api/search \
-H "Content-Type: application/json" \
-d '{"query": "python web frameworks", "mode": "discovery", "max_results": 5}'
# Multi-turn search (conversational refinement)
curl -X POST http://localhost:8000/api/search \
-H "Content-Type: application/json" \
-d '{"query": "what about async?", "session_id": "my-session"}'
# Extract content from a working URL
curl -X POST http://localhost:8000/api/extract \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com/article"}'
# Recover a dead or moved URL
curl -X POST http://localhost:8000/api/recover-url \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com/old-page", "title": "Example Article"}'
# Public health
curl http://localhost:8000/api/health
# Admin health & budgets
curl -H "Authorization: Bearer $ARGUS_ADMIN_API_KEY" \
http://localhost:8000/api/admin/health/detail
curl -H "Authorization: Bearer $ARGUS_ADMIN_API_KEY" \
http://localhost:8000/api/admin/budgets
| Mode | Use for | Example |
|---|---|---|
discovery |
Related pages, canonical sources | "Find the official docs for X" |
research |
Broad exploratory retrieval | "Latest approaches to Y?" |
recovery |
Finding moved/dead content | "This URL is 404" |
grounding |
Fact-checking with live sources | "Verify this claim about Z" |
Tier-based routing always applies first. Within each tier, the mode selects provider order.
{
"query": "python web frameworks",
"mode": "discovery",
"results": [
{"url": "https://fastapi.tiangolo.com", "title": "FastAPI", "snippet": "Modern Python web framework", "score": 0.942}
],
"total_results": 1,
"cached": false,
"traces": [
{"provider": "duckduckgo", "status": "success", "results_count": 5, "latency_ms": 312}
]
}
Each result includes url, title, snippet, domain, provider, and score. The traces array shows which providers were called and their outcomes.
{
"budgets": {
"brave": {"remaining": 1847, "monthly_usage": 153, "usage_count": 153, "exhausted": false},
"duckduckgo": {"remaining": 0, "monthly_usage": 0, "usage_count": 42, "exhausted": false}
},
"token_balances": {"jina": 9833638}
}
Each provider tracks usage. Tier 1 (monthly) uses a 30-day rolling window; tier 3 (one-time) uses a lifetime counter that never resets. When a provider hits its budget, Argus skips it and moves to the next. Free providers (SearXNG, DuckDuckGo, GitHub) have no limit. Set ARGUS_*_MONTHLY_BUDGET_USD to enforce custom limits per provider.
argus search -q "python web framework" # zero-config, uses DuckDuckGo
argus search -q "python web framework" --mode research -n 20
argus search -q "fastapi" --session my-session # multi-turn context
argus extract -u "https://example.com/article" # extract clean text
argus extract -u "https://example.com/article" -d nytimes.com # auth extraction
argus recover-url -u "https://dead.link" -t "Title"
argus health # provider status
argus budgets # budget + token balances
argus set-balance -s jina -b 9833638 # track token balance
argus test-provider -p brave # smoke-test a provider
argus serve # start API server
argus mcp serve # start MCP server
All commands support --json for structured output.
Pass session_id to any search call. Argus stores each query and extracted URL in a SQLite-backed session. Reusing the same session_id gives the broker context from prior queries — follow-up searches are automatically refined using earlier conversation context. Sessions persist across restarts. Omit session_id for stateless, one-shot searches.
Option A — Local (stdio)
Install and run on the same machine as your MCP client:
{
"mcpServers": {
"argus": {
"command": "argus",
"args": ["mcp", "serve"]
}
}
}
Use the full path if argus isn't on PATH: "/home/you/.local/bin/argus".
Works with Claude Code, Cursor, and any stdio-based MCP client.
Option B — Self-hosted server (remote clients over Tailscale)
Run Argus on one machine, connect every client over the network. No local install on clients.
On the server:
export ARGUS_API_KEY=replace-with-a-long-random-secret
argus mcp serve --transport streamable-http --host 100.126.13.70 --port 8001
On each client:
| Client | Config |
|---|---|
| Claude Code | {"mcpServers":{"argus":{"type":"sse","url":"http://<server>:8001/mcp","headers":{"Authorization":"Bearer <ARGUS_API_KEY>"}}}} |
| Antigravity | {"mcpServers":{"argus":{"serverUrl":"http://<server>:8001/mcp","headers":{"Authorization":"Bearer <ARGUS_API_KEY>"}}}} |
With Tailscale, <server> is your machine's Tailscale IP (e.g. 100.126.13.70). One server, every machine on your mesh gets search.
Transports: stdio (default, for local), sse (legacy remote), streamable-http (modern remote — required for Antigravity). Remote HTTP transports require ARGUS_API_KEY.
Available tools:
stdio: search_web, extract_content, recover_url, expand_links, search_health, search_budgets, test_provider, cookie_health, valyu_answersearch_web, extract_content, recover_url, expand_links, valyu_answerfrom argus.broker.router import create_broker
from argus.models import SearchQuery, SearchMode
from argus.extraction import extract_url
broker = create_broker()
response = await broker.search(
SearchQuery(query="python web frameworks", mode=SearchMode.DISCOVERY, max_results=10)
)
for r in response.results:
print(f"{r.title}: {r.url} (score: {r.score:.3f})")
content = await extract_url(response.results[0].url)
print(content.title)
print(content.text)
Argus tries up to ten methods to extract content from any URL: first local (trafilatura, Crawl4AI, Obscura, Playwright), then external APIs (Jina, Valyu Contents, Firecrawl, You.com, Wayback, archive.is). Each attempt is quality-checked for garbage output. See docs/providers.md for the full extractor comparison.
Completeness assessment runs automatically after every successful extraction. Argus scores five signals — trailing ellipsis, feed truncation markers ("Read more", WordPress RSS footers), mid-sentence endings, abrupt final paragraphs, and suspicious round word counts — and returns is_complete, completeness_confidence, and truncation_type alongside the text. When confidence is ≥ 85%, Argus continues trying the next extractor rather than returning a partial result; this means a trafilatura fetch that ends with "..." will automatically fall through to Playwright, Jina, Wayback, etc. Callers that already have text (e.g. RSS feed items) can use POST /api/assess-content to check completeness without triggering extraction.
Obscura (optional) is a lightweight Rust headless browser (~70MB binary, 30MB RAM) with built-in stealth mode — it sets navigator.webdriver=undefined, randomizes canvas/GPU/audio fingerprints per session, and blocks 3,520 tracker domains. This directly addresses bot detection on JS-heavy and anti-scraping sites that block standard Playwright/Chrome. No API key, no rate limit — fully local.
Two ways to use it:
| Mode | Setup | What you get |
|---|---|---|
| CLI extraction step | Install binary on $PATH |
Argus auto-detects it; stealth browser as fallback step before Playwright |
| CDP backend for Playwright | Run obscura serve --stealth --port 9222, set ARGUS_OBSCURA_CDP_URL=ws://127.0.0.1:9222 |
Playwright uses Obscura as its browser engine — stealth + 30MB vs 200MB + DOM-to-Markdown output |
Install the binary: github.com/h4ckf0r0day/obscura/releases
Extract gets the full text of a working URL and tells you whether that text is complete. Recover-URL finds alternatives when a URL is dead, paywalled, or radically changed.
Caller (CLI/HTTP/MCP/Python) → SearchBroker → tier-sorted providers → RRF ranking → response
↕ SessionStore (optional)
Extractor (on demand) → 10-step fallback chain with quality gates
| Module | Responsibility |
|---|---|
argus/broker/ |
Tier-based routing, ranking, dedup, caching, health, budgets |
argus/providers/ |
Provider adapters (one per search API) |
argus/extraction/ |
10-step URL extraction fallback chain with quality gates |
argus/sessions/ |
Multi-turn session store and query refinement |
argus/api/ |
FastAPI HTTP endpoints |
argus/cli/ |
Click CLI commands |
argus/mcp/ |
MCP server for LLM integration |
argus/persistence/ |
PostgreSQL query/result storage |
Add new providers or extractors with a single adapter file. See CONTRIBUTING.md for the interface.
query arrives → cache? → build provider queue → execute sequentially → RRF fuse → dedup → respond
Cache check. SearchCache hashes normalized_query:mode (SHA256). Hit returns immediately with a TTL of 168 hours (7 days).
Provider queue. resolve_routing() takes the mode-specific preference list and stable-sorts by tier: tier 0 (free) first, tier 1 (monthly) next, tier 3 (one-time) last. Example for discovery mode:
searxng → duckduckgo → yahoo → github → brave → exa → tavily → linkup → serper → parallel → you → valyu
Sequential execution with gates. Each provider is checked in order. Four gates must pass before an API call:
RRF fusion. Results from all queried providers are merged using Reciprocal Rank Fusion (k=60). Each result's score is the sum of 1/(k + rank) across every provider that returned it. Results appearing in multiple providers rank higher.
Dedup and truncate. URLs are normalized (stripped www., tracking params like utm_*, trailing slashes) and deduplicated. The merged list is truncated to max_results (default 10).
Cache and persist. The final response is written to the in-memory cache and persisted to PostgreSQL. Provider traces (which were called, which were skipped and why) are included in the response for observability.
All config via environment variables. See .env.example for the full list. Missing keys degrade gracefully — providers are skipped, not errors.
| Variable | Default | Description |
|---|---|---|
ARGUS_SEARXNG_BASE_URL |
http://127.0.0.1:8080 |
SearXNG endpoint |
ARGUS_BRAVE_API_KEY |
— | Brave Search API key |
ARGUS_SERPER_API_KEY |
— | Serper API key |
ARGUS_TAVILY_API_KEY |
— | Tavily API key |
ARGUS_EXA_API_KEY |
— | Exa API key |
ARGUS_LINKUP_API_KEY |
— | Linkup API key |
ARGUS_PARALLEL_API_KEY |
— | Parallel AI API key |
ARGUS_YOU_API_KEY |
— | You.com API key |
ARGUS_VALYU_API_KEY |
— | Valyu API key (search, contents, answer) |
ARGUS_FIRECRAWL_API_KEY |
— | Firecrawl API key (content extraction) |
ARGUS_GITHUB_API_KEY |
— | GitHub token (higher rate limit) |
ARGUS_*_MONTHLY_BUDGET_USD |
0 (unlimited) | Query-count budget per provider |
ARGUS_CRAWL4AI_ENABLED |
false | Enable Crawl4AI extraction step |
ARGUS_YOU_CONTENTS_ENABLED |
false | Enable You.com Contents API extraction |
ARGUS_OBSCURA_CDP_URL |
— | Obscura CDP endpoint (e.g. ws://127.0.0.1:9222) — makes Playwright use Obscura as its browser engine |
ARGUS_OBSCURA_TIMEOUT_SECONDS |
20 | Timeout for Obscura CLI subprocess calls |
ARGUS_CACHE_TTL_HOURS |
168 | Result cache TTL |
How is this different from calling Tavily/Serper directly? Argus calls them for you — plus 13 other providers. You get one ranked, deduplicated result set instead of managing multiple API keys and stitching results together. Free providers are tried first, so you only burn credits when needed.
Can I run only one provider? Yes. Set only the API key for the provider you want. All others are silently skipped. For zero-config, just install and go — DuckDuckGo + Yahoo handle search with no keys.
Do I need Docker?
No. pip install argus-search works immediately on any machine with Python 3.11+. Docker is only needed for SearXNG (self-hosted search, aggregates 70+ engines) or Crawl4AI (local JS rendering).
Which Python version should contributors use?
Use Python 3.12 for repo development and verification: uv sync --python 3.12 --extra dev --extra mcp then uv run pytest tests/ -v --tb=short. The published package still supports Python 3.11+.
What is the safest way to deploy Argus on a network?
Use Tailscale or another private network, bind explicitly to the trusted interface, set ARGUS_API_KEY, and reserve /api/admin/* for ARGUS_ADMIN_API_KEY. Treat direct internet exposure as an advanced mode behind a reverse proxy.
MIT — see CHANGELOG.md for release history.
Добавь это в claude_desktop_config.json и перезапусти Claude Desktop.
{
"mcpServers": {
"khamel83-argus": {
"command": "npx",
"args": []
}
}
}