ymylive

free-search-mcp

Community ymylive
Updated

Local-first, no-API-key MCP server for web search, fetch, and document reading. Multi-engine (DuckDuckGo, Mojeek, Startpage), smart Playwright fallback, FTS5 cache, LLM-tuned output.

free-search-mcp

LicensePythonMCP

A local-first, no-API-key Model Context Protocol server that gives anyLLM (Claude, GPT, local Ollama, …) the ability to search the web, fetch andclean up pages, and read documents — without you signing up for a singlesearch API.

It bundles together the best ideas from a handful of open-source MCPs intoone Python package, and adds the LLM-ergonomics and reliability work theywere each missing.

research("how does reciprocal rank fusion work", depth=3)
   ↓
# Research brief: how does reciprocal rank fusion work
_engines: duckduckgo, mojeek, startpage · sources: 3 · ~3,400 tokens_

## Sources
- [1] Reciprocal rank fusion | Elasticsearch Reference — <https://…>
- [2] Hybrid Search Scoring (RRF) | Microsoft Learn — <https://…>
- [3] RRF explained in 4 mins — Medium — <https://…>

## Documents
…full Markdown bodies of each page, ready for the LLM to read…

One tool call. Three sources. No API key. No OPENAI_API_KEY-but-for-searchshakedown.

Why this exists

Existing search MCPs each do one thing well, but you usually want all of it:

Multi-engine No API key Smart fallback PDF/DOCX FTS5 cache Filters Trafilatura LLM-tuned
nickclyde/duckduckgo-mcp-server ~
mrkrsl/web-search-mcp ~
Aas-ee/open-webSearch ~ ~
VincentKaufmann/noapi-google-search-mcp ~
free-search-mcp

"LLM-tuned" here means: Markdown-first output, token estimates, smarttruncation at paragraph boundaries, "Best for / Not for / Returns / Commonmistakes" docstrings the model uses to pick the right tool, actionableerror hints, MCP prompts and resource templates, and a one-shotresearch() that collapses search→fetch→fetch→fetch into a single turn.

"Trafilatura" means we extract main content usingtrafilatura — winner of theBevendorff 2023 ROUGE benchmark (~0.85 vs ~0.55 for naive boilerplatestripping). Each fetched page also returns author, published_date, andsitename for free.

"Filters" means search/research accept freshness, include_domains,exclude_domains, category (news/pdf/github/paper/forum/blog),include_text, exclude_text.

Tools

Tool Description
search(query, engines?, max_results?, use_cache?, max_age_hours?, freshness?, include_domains?, exclude_domains?, category?, include_text?, exclude_text?, format?) Parallel multi-engine search merged via Reciprocal Rank Fusion
research(question, depth?, engines?, fetch?, use_cache?, max_age_hours?, freshness?, include_domains?, exclude_domains?, category?, include_text?, exclude_text?, format?) One-shot: search + fetch top N + return Markdown brief
fetch(url, render?, force_refresh?, max_age_hours?, format?) Fetch a page, return reader-mode Markdown (trafilatura-extracted, with author/date/sitename)
fetch_batch(urls, render?, format?) Concurrent multi-URL fetch
read_doc(source, start?, length?, format?) Parse PDF / DOCX / HTML / TXT / MD with pagination
cache_search(query, limit?, format?) FTS5 search across previously fetched pages
engines() List engine names available to search

Plus 2 MCP prompts (Research thoroughly, Fact-check claim) and aresource template (cache://page/{url}) for dragging cached pages backinto context without re-fetching.

Filters (search / research)

Param Values Effect
freshness day / week / month / year Only results from the last N
include_domains ["python.org", "djangoproject.com"] Restrict to these domains
exclude_domains ["pinterest.com"] Remove these
category news / pdf / github / paper / forum / blog Content-type shortcut (paper = arxiv/acm/ieee/…, forum = reddit/HN/SE, etc.)
include_text "async" Substring required in title/snippet
exclude_text "beginner" Substring forbidden
max_age_hours 24 Override the 7-day default cache TTL on this call

All tools default to format="markdown" — readable, ~40% fewer tokens thanJSON, with provenance and a token-budget header. Pass format="json" forstructured access.

Tool annotations

Every tool ships correct readOnlyHint, idempotentHint, andopenWorldHint annotations so MCP clients can label them and gateelevated actions.

Engines

Default set (all reliable, no captchas during repeated calls):duckduckgo, mojeek, startpage.

Opt-in (intermittent challenges to headless clients):brave, bing, baidu.

Brave/Bing/Baidu all gate headless browsers after a handful of calls (PoWCAPTCHAs, "something went wrong" pages, redirect wrappers). Passengines=["brave"] etc. only when the defaults can't find what you need.

Install

git clone https://github.com/ymylive/free-search-mcp.git
cd free-search-mcp
uv sync
uv run playwright install chromium

Run as a stand-alone server (stdio transport):

uv run search-mcp

Run live tests (hits the real web — set the env var):

SEARCH_MCP_TEST_NETWORK=1 uv run pytest -v

Offline tests run by default and don't touch the network.

Wire into Claude Desktop

Add this to ~/Library/Application Support/Claude/claude_desktop_config.json(macOS) or the equivalent on your platform:

{
  "mcpServers": {
    "search": {
      "command": "uv",
      "args": ["--directory", "/absolute/path/to/free-search-mcp", "run", "search-mcp"]
    }
  }
}

Restart Claude Desktop. The seven tools above will appear in the tooldrawer.

Wire into other clients

The server speaks plain MCP over stdio. Anything that supports MCP works:

  • Claude Code (claude mcp add search uv --directory /…/free-search-mcp run search-mcp)
  • Cursor / Continue / Cline (use the JSON snippet above)
  • Custom Python / TypeScript clients via the official MCP SDK

Configuration

All settings can be overridden by environment variables prefixed withSEARCH_MCP_:

Var Default Meaning
SEARCH_MCP_DEFAULT_ENGINES ["duckduckgo","mojeek","startpage"] JSON list
SEARCH_MCP_MAX_RESULTS_PER_ENGINE 10
SEARCH_MCP_RATE_LIMIT_PER_MINUTE 30 per engine
SEARCH_MCP_FETCH_RATE_LIMIT_PER_MINUTE 20 shared fetch bucket
SEARCH_MCP_CACHE_DIR ~/.cache/search-mcp
SEARCH_MCP_CACHE_TTL_SECONDS 604800 7 days
SEARCH_MCP_FETCH_STRATEGY auto auto / http / browser
SEARCH_MCP_BROWSER_HEADLESS true
SEARCH_MCP_BROWSER_POOL_SIZE 2 concurrent pages
SEARCH_MCP_MAX_CONTENT_CHARS 50000 per result truncation

Architecture

   ┌─────────────────────────────────────────────────────┐
   │  FastMCP server (stdio)                             │
   │  tools: search / research / fetch / fetch_batch /   │
   │         read_doc / cache_search / engines           │
   └────────────┬────────────────────────────────────────┘
                │
   ┌────────────▼────────────┐  ┌────────────────────────┐
   │  aggregator             │  │  fetcher               │
   │  - parallel engines     │  │  - httpx fast path     │
   │  - reciprocal rank      │  │  - playwright fallback │
   │    fusion               │  │  - markdownify         │
   │  - search cache (FTS5)  │  │  - page cache (FTS5)   │
   └────┬────────────────────┘  └────────────┬───────────┘
        │                                    │
   ┌────▼─────────────────┐  ┌──────────────▼─────────────┐
   │  engines/            │  │  browser pool              │
   │   duckduckgo.py      │  │   - persistent context     │
   │   mojeek.py          │  │   - stealth init script    │
   │   startpage.py       │  │   - shared cookies         │
   │   brave.py     (opt) │  │   - semaphore-bounded pages│
   │   bing.py      (opt) │  └────────────────────────────┘
   │   baidu.py     (opt) │
   └──────────────────────┘

   ┌────────────────────────────┐    ┌──────────────────┐
   │  documents/                │    │  ratelimit       │
   │   pypdf, python-docx,      │    │   token bucket   │
   │   markdownify              │    │   per engine     │
   └────────────────────────────┘    └──────────────────┘

   ┌────────────────────────────┐    ┌──────────────────┐
   │  formatting                │    │  research        │
   │   token estimate           │    │   composed       │
   │   smart truncation         │    │   workflow       │
   │   markdown renderers       │    │                  │
   └────────────────────────────┘    └──────────────────┘

Engine adapter pattern

Each engine in src/search_mcp/engines/ implements:

class Engine:
    name: str
    needs_browser: bool          # Force Playwright?
    wait_selector: str | None    # CSS to wait for in browser mode

    def build_url(self, query: str, max_results: int) -> str: ...
    def parse(self, html: str) -> list[SearchResult]: ...

The base class handles transport (httpx → Playwright fallback), ratelimiting, and the case where HTTP returns a captcha shell instead ofresults (auto-retries via the browser).

Credits

This project stands on the shoulders of:

License

MIT — see LICENSE.

MCP Server · Populars

MCP Server · New