Provides a bridge between Ollama and the Model Context Protocol, enabling access to Ollama's local LLM capabilities including model management (pull, push, list, create), model execution with customizable parameters, vision/multimodal support, and advanced reasoning via the 'think' parameter.
Offers an OpenAI-compatible chat completion API interface, allowing the server to function as a drop-in replacement for OpenAI's chat completion functionality while using Ollama's local LLM models.
Ollama MCP Server
This is a rebooted and actively maintained fork.
Original project: NightTrek/Ollama-mcpThis repository (hyzhak/ollama-mcp-server) is a fresh upstream with improved maintenance, metadata, and publishing automation.
See NightTrek/Ollama-mcp for project history and prior releases.
🚀 A powerful bridge between Ollama and the Model Context Protocol (MCP), enabling seamless integration of Ollama's local LLM capabilities into your MCP-powered applications.
🌟 Features
Complete Ollama Integration
- Full API Coverage: Access all essential Ollama functionality through a clean MCP interface
- OpenAI-Compatible Chat: Drop-in replacement for OpenAI's chat completion API
- Local LLM Power: Run AI models locally with full control and privacy
Core Capabilities
- 🔄 Model Management
- Pull models from registries
- Push models to registries
- List available models
- Create custom models from Modelfiles
- Copy and remove models
- 🤖 Model Execution
- Run models with customizable prompts (response is returned only after completion; streaming is not supported in stdio mode)
- Vision/multimodal support: pass images to compatible models
- Chat completion API with system/user/assistant roles
- Configurable parameters (temperature, timeout)
- NEW:
think
parameter for advanced reasoning and transparency (see below) - Raw mode support for direct responses
- 🛠 Server Control
- Start and manage Ollama server
- View detailed model information
- Error handling and timeout management
🚀 Quick Start
Prerequisites
Configuration
Add the server to your MCP configuration:
For Claude Desktop:
MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
Windows: %APPDATA%/Claude/claude_desktop_config.json
🛠 Developer Setup
Prerequisites
- Ollama installed on your system
- Node.js and npm
Installation
- Install dependencies:
- Build the server:
🛠 Usage Examples
Pull and Run a Model
Run a Vision/Multimodal Model
Chat Completion (OpenAI-compatible)
Note: The
images
field is optional and only supported by vision/multimodal models.
Create Custom Model
🧠 Advanced Reasoning with the think
Parameter
Both the run
and chat_completion
tools now support an optional think
parameter:
think: true
: Requests the model to provide step-by-step reasoning or "thought process" in addition to the final answer (if supported by the model).think: false
(default): Only the final answer is returned.
Example (run
tool):
- If the model supports it, the response will include a
:::thinking ... :::
block with detailed reasoning before the final answer.
Example (chat_completion
tool):
- The model's reasoning (if provided) will be included in the message content.
Note: Not all models support the
think
parameter. Advanced models (e.g., "deepseek-r1:32b", "magistral") may provide more detailed and accurate reasoning whenthink
is enabled.
🔧 Advanced Configuration
OLLAMA_HOST
: Configure custom Ollama API endpoint (default: http://127.0.0.1:11434)- Timeout settings for model execution (default: 60 seconds)
- Temperature control for response randomness (0-2 range)
🤝 Contributing
Contributions are welcome! Feel free to:
- Report bugs
- Suggest new features
- Submit pull requests
📝 License
MIT License - feel free to use in your own projects!
Built with ❤️ for the MCP ecosystem
This server cannot be installed
local-only server
The server can only run on the client's local machine because it depends on local resources.
A bridge that integrates Ollama's local LLM capabilities into MCP-powered applications, enabling users to run, manage, and interact with AI models locally with full control and privacy.
Related MCP Servers
- -securityFlicense-qualityAn interactive chat interface that combines Ollama's LLM capabilities with PostgreSQL database access through the Model Context Protocol (MCP). Ask questions about your data in natural language and get AI-powered responses backed by real SQL queries.Last updated -45TypeScript
- AsecurityAlicenseAqualityMCP Ollama server integrates Ollama models with MCP clients, allowing users to list models, get detailed information, and interact with them through questions.Last updated -312PythonMIT License
- -securityAlicense-qualityEnables seamless integration between Ollama's local LLM models and MCP-compatible applications, supporting model management and chat interactions.Last updated -9987TypeScriptAGPL 3.0
- AsecurityFlicenseAqualityA bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.Last updated -1033JavaScript