deficlow

HyperStore MCP

Community deficlow
Updated

HyperStore-MCP

HyperStore MCP

Plug 6,500+ AI apps into any LLM via the Model Context Protocol.

PyPICILicense: MIT

HyperStore is a curated directory of 6,500+ AI applications, developed by HyperGPT.This MCP server exposes the HyperStore catalog to any LLM client — Claude, ChatGPT, Cursor,Windsurf, Cline, Zed, Gemini, and anything else that speaks MCP.

Ask your LLM:

"Find me a free AI tool that summarises PDFs.""Compare ChatGPT, Claude, and Gemini side-by-side.""Show me the top 5 image-generation apps with an API."

The LLM calls HyperStore MCP behind the scenes and answers with up-to-date, curated results.

What you get

8 tools:

Tool Purpose
search_apps Full-text keyword search
ai_search Embedding-based semantic search
get_app Full app detail (features, screenshots, pricing)
list_apps Paginated apps with filters (category, pricing)
list_categories Browse all 30+ categories
category_apps Apps within a category
browse_apps A-Z directory listing
get_homepage Trending + top categories overview

3 resources:

  • hyperstore://app/{slug} — markdown rendering of any app
  • hyperstore://category/{slug} — top apps in a category
  • hyperstore://catalog — full category index

3 prompts:

  • find_tool_for_task — guided discovery for a task
  • compare_apps — side-by-side app comparison
  • discover_category — explore a topic

Install

Option A — uvx (zero install, recommended)

Requires uv. One command and you're done:

uvx hyperstore-mcp

Option B — pipx

pipx install hyperstore-mcp
hyperstore-mcp

Option C — Docker (for remote hosting)

docker run --rm -p 8080:8080 ghcr.io/deficlow/hyperstore-mcp
# Now MCP Streamable HTTP at http://localhost:8080/mcp

Option D — Hosted endpoint (no install)

Use our managed Streamable HTTP server:

https://mcp.store.hypergpt.ai/mcp

Connect from your LLM client

Claude Desktop

Edit ~/Library/Application Support/Claude/claude_desktop_config.json(macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows):

{
  "mcpServers": {
    "hyperstore": {
      "command": "uvx",
      "args": ["hyperstore-mcp"]
    }
  }
}

Restart Claude → tools appear in the 🛠 menu.

Claude Code

claude mcp add hyperstore -- uvx hyperstore-mcp

Cursor

.cursor/mcp.json (project) or ~/.cursor/mcp.json (global):

{
  "mcpServers": {
    "hyperstore": {
      "command": "uvx",
      "args": ["hyperstore-mcp"]
    }
  }
}

Windsurf

~/.codeium/windsurf/mcp_config.json:

{
  "mcpServers": {
    "hyperstore": {
      "command": "uvx",
      "args": ["hyperstore-mcp"]
    }
  }
}

Cline (VS Code)

settings.json:

{
  "cline.mcpServers": {
    "hyperstore": {
      "command": "uvx",
      "args": ["hyperstore-mcp"]
    }
  }
}

Zed

~/.config/zed/settings.json:

{
  "context_servers": {
    "hyperstore": {
      "command": {
        "path": "uvx",
        "args": ["hyperstore-mcp"]
      }
    }
  }
}

Gemini CLI

~/.gemini/settings.json:

{
  "mcpServers": {
    "hyperstore": {
      "command": "uvx",
      "args": ["hyperstore-mcp"]
    }
  }
}

ChatGPT (Pro / Team / Enterprise)

Settings → Connectors → Add custom connector:

  • Name: HyperStore
  • MCP Server URL: https://mcp.store.hypergpt.ai/mcp
  • Authentication: None

OpenAI Responses API

from openai import OpenAI

client = OpenAI()
response = client.responses.create(
    model="gpt-4.1",
    tools=[{
        "type": "mcp",
        "server_label": "hyperstore",
        "server_url": "https://mcp.store.hypergpt.ai/mcp",
        "require_approval": "never",
    }],
    input="Find me 3 free AI tools for writing unit tests.",
)
print(response.output_text)

Anthropic Messages API

from anthropic import Anthropic

client = Anthropic()
response = client.messages.create(
    model="claude-opus-4-7",
    max_tokens=1024,
    mcp_servers=[{
        "type": "url",
        "url": "https://mcp.store.hypergpt.ai/mcp",
        "name": "hyperstore",
    }],
    messages=[{"role": "user", "content": "Top 5 AI image generators?"}],
)

See examples/ for ready-to-paste configs for every supported client.

Run as a remote server

# Streamable HTTP (modern, ChatGPT/OpenAI/Anthropic)
hyperstore-mcp --transport http --host 0.0.0.0 --port 8080

# Legacy SSE (older MCP clients)
hyperstore-mcp --transport sse --port 8080

The hosted endpoint at https://mcp.store.hypergpt.ai runs the Docker imagebehind a CDN — no auth, rate-limited per IP.

Configuration

All settings come from environment variables (see .env.example):

Variable Default Purpose
HYPERSTORE_API_BASE https://store.hypergpt.ai Upstream API base URL
HYPERSTORE_TIMEOUT 20 HTTP timeout in seconds
HYPERSTORE_USER_AGENT hyperstore-mcp/{version} UA string
MCP_HOST 0.0.0.0 Bind host (http/sse only)
MCP_PORT 8080 Bind port (http/sse only)
LOG_LEVEL INFO Logging level

Development

git clone https://github.com/deficlow/HyperStore-MCP
cd HyperStore-MCP
uv sync --all-extras
uv run pytest
uv run hyperstore-mcp        # stdio mode for local testing

Inspect the running server with the official MCP Inspector:

npx @modelcontextprotocol/inspector uvx hyperstore-mcp

How it works

HyperStore MCP is a thin async wrapper around the HyperStore publicREST API. It is read-only — no credentials, no writes, no PII. The same data thatpowers the website powers the MCP server. Updates land in your LLM the moment theyland on the site.

LLM client ──MCP──▶ hyperstore-mcp ──HTTPS──▶ store.hypergpt.ai/api

License

MIT © HyperGPT

MCP Server · Populars

MCP Server · New

    uarlouski

    🚀 TestRail MCP Server

    AI-native MCP server connecting Claude, Cursor, Windsurf, and other AI assistants to TestRail — manage test cases, runs, and results through natural-language conversation, with typed schemas built for LLMs.

    Community uarlouski
    metabase

    Metabase MCP Server

    The easy-to-use open source Business Intelligence and Embedded Analytics tool that lets everyone work with data :bar_chart:

    Community metabase
    mindsdb

    USE CASES

    Platform dedicated to building an open foundation for applied Artificial Intelligence, designed for people seeking production-ready AI systems they can truly control, extend and deploy anywhere.

    Community mindsdb
    reflex-search

    Reflex

    Reflex - The instant, code-aware local search engine.

    Community reflex-search
    Licinexus

    @licinexusbr/mcp

    MCP server for Brazilian public procurement data (PNCP + Receita Federal). Maintained by Licinexus.

    Community Licinexus