Integrates Google's Gemini models (gemini-2.0-flash-exp, gemini-1.5-pro) for AI generation with automatic model selection based on task complexity
Supports local AI models through Ollama integration with custom API endpoints for self-hosted AI generation
Provides access to OpenAI models including GPT-4o, GPT-4o-mini, O1, and O3-mini with intelligent model selection and cost control features
Uses Redis for conversation memory management and persistent storage of conversation context across multi-turn interactions
🧙 SAGE-MCP: Simple AI Guidance Engine for Claude
Universal AI assistant MCP server with intelligent mode selection, conversation continuity, and smart file handling
SAGE-MCP transforms Claude into a multi-talented development assistant that adapts to your needs. Whether you're debugging code, planning architecture, writing tests, or having a technical discussion, SAGE automatically selects the right approach and model for optimal results.
✨ Key Features
🎯 Intelligent Mode System
chat - Natural conversations with context awareness
analyze - Deep code analysis and pattern recognition
review - Comprehensive code reviews with actionable feedback
debug - Systematic debugging and root cause analysis
plan - Strategic project planning and architecture design
test - Test generation with coverage analysis
refactor - Code improvement and modernization
think - Deep reasoning with adjustable thinking depth
🔄 Conversation Continuity
Seamless multi-turn conversations across different modes
Automatic context preservation between tool calls
Smart file deduplication - never re-read the same files
Thread-based memory system for long-running tasks
🤖 Smart Model Selection
Auto mode - Intelligent model selection based on task complexity
Support for multiple providers: OpenAI, Anthropic, Google, OpenRouter
Model restrictions via environment variables for cost control
Thinking depth control: minimal (0.5%), low (8%), medium (33%), high (67%), max (100%)
📁 Intelligent File Handling
embedded - Full file content in context (default)
summary - Token-efficient summaries for large codebases
reference - File storage with ID references
Automatic directory expansion and smart deduplication
Security validation for all file operations
🌐 Web Search Integration
Real-time documentation lookup
Best practices and current standards
Framework and library research
Error and issue investigation
🎨 Mode Specializations
Mode | Temperature | Description | Best For |
chat | 0.5 | Natural conversations with balanced creativity | Q&A, brainstorming, explanations |
analyze | 0.2 | Focused precision for code analysis | Architecture review, pattern detection |
review | 0.3 | Systematic evaluation with consistent standards | Security audits, best practices |
debug | 0.1 | Deterministic analysis for troubleshooting | Error investigation, root cause analysis |
plan | 0.4 | Strategic thinking for project planning | Architecture design, task breakdown |
test | 0.2 | Accurate test generation with edge cases | Unit tests, integration tests |
refactor | 0.3 | Careful improvements preserving functionality | Code modernization, optimization |
think | 0.7 | Creative problem solving with deep reasoning | Complex algorithms, system design |
🚀 Quick Start
Installation
Claude Desktop Configuration
Add to your Claude Desktop MCP settings:
📖 Usage Examples
Basic Chat
Code Analysis with Files
Multi-turn Conversations
Deep Thinking Mode
Smart File Handling
⚙️ Configuration
Environment Variables
Mode-Specific Temperatures
Default temperatures optimized for each mode:
chat: 0.5 - Balanced creativity
analyze: 0.2 - Focused precision
review: 0.3 - Systematic evaluation
debug: 0.1 - Deterministic analysis
plan: 0.4 - Strategic thinking
test: 0.2 - Accurate test generation
refactor: 0.3 - Careful improvements
think: 0.7 - Creative problem solving
🔧 Advanced Features
Conversation Continuation
Smart File Handling
Model Restrictions
Supported Models
Provider | Models | Configuration |
OpenAI | gpt-4o, gpt-4o-mini, o1, o3-mini | OPENAI_API_KEY |
Anthropic | claude-3-5-sonnet, claude-3-5-haiku | ANTHROPIC_API_KEY |
gemini-2.0-flash-exp, gemini-1.5-pro | GOOGLE_API_KEY | |
OpenRouter | 100+ models from all providers | OPENROUTER_API_KEY |
Custom/Ollama | llama3.2, mistral, codestral | CUSTOM_API_URL |
Complete Configuration Reference
Variable | Description | Example |
API Keys | ||
| OpenAI API key |
|
| Anthropic Claude API key |
|
/
| Google Gemini API key |
|
| OpenRouter API key |
|
| xAI (Grok) API key |
|
| Custom/Ollama API endpoint |
|
| Custom API key (if required) |
|
Model Selection | ||
| Default model (
for selection) |
,
,
|
Model Restrictions | ||
| Allowed OpenAI models |
|
| Allowed Google models |
|
| Allowed Anthropic models |
|
| Blocked models (any provider) |
|
| Disable by pattern |
|
Limits & Performance | ||
| Maximum file size in bytes |
(5MB) |
| MCP transport limit |
|
| Max turns per conversation |
|
| Conversation timeout |
|
Memory & Storage | ||
| Redis connection for memory |
|
| Redis database number |
|
Temperature Overrides | ||
| Chat mode temperature |
|
| Analyze mode temperature |
|
| Debug mode temperature |
|
| Plan mode temperature |
|
| Test mode temperature |
|
| Refactor mode temperature |
|
| Review mode temperature |
|
| Think mode temperature |
|
🏗️ Architecture
🧪 Advanced Features
Model Restrictions
Control which models can be used to manage costs:
Conversation Memory
SAGE maintains conversation context across tool calls:
Custom Providers
Add custom AI providers by implementing the base provider interface:
🤝 Contributing
We welcome contributions! Please see CONTRIBUTING.md for guidelines.
Development Setup
📄 License
MIT License - see LICENSE for details.
🙏 Acknowledgments
Built on FastMCP framework
Inspired by zen-mcp-server
Powered by Claude MCP protocol
🔗 Links
SAGE-MCP - Your intelligent AI assistant that adapts to how you work 🧙✨
This server cannot be installed
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
A universal AI assistant MCP server that transforms Claude into a multi-talented development assistant with intelligent mode selection, conversation continuity, and smart file handling. Automatically adapts to different tasks like debugging, code analysis, planning, and testing while supporting multiple AI providers and maintaining context across conversations.
Related MCP Servers
- AsecurityAlicenseAqualityAn MCP server that implements Claude Code-like functionality, allowing the AI to analyze codebases, modify files, execute commands, and manage projects through direct file system interactions.Last updated -15265MIT License
- -securityFlicense-qualityAn MCP server that allows AI assistants like Claude to execute terminal commands on the user's computer and return the output, functioning like a terminal through AI.Last updated -57
- AsecurityAlicenseAqualityMCP server that provides Claude AI assistants with the ability to search the web, get news, and perform research using the You.com API.Last updated -41MIT License
- -securityFlicense-qualityA production-ready MCP server ecosystem providing Claude AI with 150+ specialized tools across enhanced memory, data analytics, security, design, and infrastructure domains with PostgreSQL, Redis, Qdrant, and Docker orchestration.Last updated -