Skip to main content
Glama

LangExtract MCP Server

by larsenweigle
supported-models.md2.75 kB
# Supported Language Models This document provides comprehensive information about the language models supported by the langextract-mcp server. ## Currently Supported Models The langextract-mcp server currently supports **Google Gemini models only**, which are optimized for reliable structured extraction with schema constraints. ### Gemini 2.5 Flash - **Provider**: Google - **Model ID**: `gemini-2.5-flash` - **Description**: Fast, cost-effective model with excellent quality - **Schema Constraints**: ✅ Supported - **Recommended For**: - General extraction tasks - Fast processing requirements - Cost-sensitive applications - **Notes**: Recommended default choice - optimal balance of speed, cost, and quality ### Gemini 2.5 Pro - **Provider**: Google - **Model ID**: `gemini-2.5-pro` - **Description**: Advanced model for complex reasoning tasks - **Schema Constraints**: ✅ Supported - **Recommended For**: - Complex extractions - High accuracy requirements - Sophisticated reasoning tasks - **Notes**: Best quality for complex tasks but higher cost ## Model Recommendations | Use Case | Recommended Model | Reason | |----------|------------------|---------| | **Default/General** | `gemini-2.5-flash` | Best balance of speed, cost, and quality | | **High Quality** | `gemini-2.5-pro` | Superior accuracy and reasoning capabilities | | **Cost Optimized** | `gemini-2.5-flash` | Most cost-effective option | | **Complex Reasoning** | `gemini-2.5-pro` | Advanced reasoning for complex extraction tasks | ## Configuration Parameters When using any supported model, you can configure the following parameters: - **`model_id`**: The model identifier (e.g., "gemini-2.5-flash") - **`max_char_buffer`**: Maximum characters per chunk (default: 1000) - **`temperature`**: Sampling temperature 0.0-1.0 (default: 0.5) - **`extraction_passes`**: Number of extraction passes for better recall (default: 1) - **`max_workers`**: Maximum parallel workers (default: 10) ## Limitations - **Provider Support**: Currently supports Google Gemini models only - **Future Support**: OpenAI and local model support may be added in future versions - **API Dependencies**: Requires active internet connection and valid API keys ## Schema Constraints All supported Gemini models include schema constraint capabilities, which means: - **Structured Output**: Guaranteed JSON structure based on your examples - **Type Safety**: Consistent field types across extractions - **Validation**: Automatic validation of extracted data against schema - **Reliability**: Reduced hallucination and improved consistency This makes the langextract-mcp server particularly reliable for production applications requiring consistent structured data extraction.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/larsenweigle/langextract-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server