๐ MCP Stock Analyzer
An official Model Context Protocol (MCP) server for dynamically analyzing Global/Indian stocks and offline local NSE BhavData documents using AI.
โญ Core Capabilities:
- Global Stocks: Automatically grabs live prices, history, fundamentals, and ticker resolution via
yfinance. - Local BhavData: AI dynamically writes SQL statements to extract massive local datasets.
- โ ๏ธ CRITICAL WARNING: Do not use the "Add Context" button (or drag-and-drop) to upload huge BhavData CSVs directly into your chat. This will instantly blow up your AI's token limits and cost a fortune. Instead, simply paste the absolute file path (e.g.,
"Analyze C:\downloads\bhav.csv") as normal text in your prompt, and let this MCP server securely query it in the background!
- โ ๏ธ CRITICAL WARNING: Do not use the "Add Context" button (or drag-and-drop) to upload huge BhavData CSVs directly into your chat. This will instantly blow up your AI's token limits and cost a fortune. Instead, simply paste the absolute file path (e.g.,
- Visual Dashboards: The AI creates completely interactive, localized HTML graphing dashboards on demand.
๐๏ธ System Architecture
graph TD
User([๐ค End User])
LLM_Client[๐ง AI Client / IDE <br> e.g., Cline, Claude, Ollama]
subgraph "Python MCP Server (.venv)"
Server[๐ server.py <br> FastMCP Entrypoint]
GlobalMod[๐ global_stocks.py <br> yfinance API]
BhavMod[๐ bhavdata_analyzer.py <br> SQLite & Pandas]
DashMod[๐ dashboard_generator.py <br> Chart.js & HTML]
end
subgraph "External Providers"
YFinance[(Yahoo Finance API)]
LocalDisk[(User's Local C: Drive <br> .csv files)]
CDN[(Chart.js CDN)]
end
User -->|Sends Prompt & Files| LLM_Client
LLM_Client -->|Calls JSON Tools via stdio| Server
Server -->|Routes query| GlobalMod
Server -->|Routes query| BhavMod
Server -->|Routes query| DashMod
GlobalMod <-->|Fetches real-time/historical data| YFinance
BhavMod <-->|Loads/Runs SQL on| LocalDisk
DashMod -->|Embeds| CDN
DashMod -->|Outputs temp HTML file to| LocalDisk
Server -.->|Returns result context| LLM_Client
LLM_Client -.->|Streams final answer to| User
๐ ๏ธ Installation & Quick Setup
To guarantee there are zero package or import errors, please set up the isolated environment:
- Open a terminal navigating to the project folder (
d:\Projects\MCPAgentStockAnalyzer). - Run these configuration commands to set up the dependencies firmly in a
.venv:python -m venv .venv .\.venv\Scripts\activate pip install -r requirements.txt
(If you use VSCode, .vscode/settings.json is automatically pre-configured to select this virtual environment.)
๐ Connecting to LLM Clients
To have your AI interact with this server, you'll simply embed its configuration into your specific tool's config file.
Here is the master Configuration JSON you will use for every client listed below:
"StockAnalyzer": {
"command": "d:/Projects/MCPAgentStockAnalyzer/.venv/Scripts/python.exe",
"args": ["d:/Projects/MCPAgentStockAnalyzer/src/server.py"]
}
1. Claude Desktop (Mac / Windows)
- Open up Claude Desktop application.
- Go to Settings > Settings file or navigate to
%APPDATA%\Claude\claude_desktop_config.json. - Add the
StockAnalyzerblock above right inside the"mcpServers": { ... }object. - Restart Claude Desktop.
2. Cline (VS Code Extension)
- In VS Code, click the Cline Extensions icon on the sidebar.
- Click the specific MCP server (plugin) settings config near the bottom.
- Paste the configuration block directly into your
mcp_settings.json.
3. Antigravity (Local IDE Agent)
- Inside your
~/.gemini/antigravity/folder (or active brain/project.geminifolder). - Edit the
mcp_config.jsonfile. - Drop the
StockAnalyzerblock into"mcpServers". Keep chatting and it hot-reloads dynamically!
4. GitHub Copilot
Currently, GitHub Copilot integrates officially directly with the Claude or OpenAI engines on newer IDE builds via specific marketplace extensions. If utilizing Copilot Chat, ensure you rely on an editor like VSCode or Cursor equipped directly with extensible Tool/Plugin settings (similar to Cline's mcp_settings.json) that bridge custom MCP standard definitions.
5. Claude Code (CLI)
If you're using Anthropic's new direct CLI tool (claude-code), configure it simply by defining it as a server in its explicit config file:
claude config set --mcp-server StockAnalyzer "d:/Projects/MCPAgentStockAnalyzer/.venv/Scripts/python.exe d:/Projects/MCPAgentStockAnalyzer/src/server.py"
๐ฆ Running completely FREE (Local LLMs)
You do not need a paid Claude 3.5 Sonnet or OpenAI API key to use this. You can direct local frameworks through Cline or Cursor directly to a local engine.
Using Ollama
- Download Ollama
- Open terminal and run a fast coder model:
ollama run qwen2.5-coder:7b - Point Cline's settings to Base URL:
http://localhost:11434/v1
Using LM Studio
- Download LM Studio
- Load any GGUF model (
Llama-3.1-8B-Instruct) and start the Local Web Server plugin (Port 1234). - Set your Cline settings provider to OpenAI Compatible, Base URL:
http://localhost:1234/v1.
โจ How to Trigger the Dashboard Tool
Simply ask your LLM: "Show me the graphical performance of Reliance from the past 6 months."The MCP tool will instantaneously compile a stunning Chart.js dashboard, save it explicitly to your temporary disk directory, and give you the local URL (e.g., file:///C:/temp/RELIANCE_dashboard_6mo.html) to tap directly in your Chrome browser!