llm-inference-mcp

zhaohongyuziranerran/llm-inference-mcp
★ 0 stars Python 🤖 AI/LLM Updated today
LLM Inference MCP Server - Multi-model routing, cost optimization, structured output. Supports OpenAI, DeepSeek, Anthropic, vLLM.
View on GitHub →

Quick Install

Copy the config for your editor. Some servers may need additional setup — check the README.

Add to claude_desktop_config.json:

{
  "mcpServers": {
    "llm-inference-mcp": {
      "command": "uvx",
      "args": [
        "llm-inference-mcp"
      ]
    }
  }
}

Or install with pip: pip install llm-inference-mcp