Skip to main content
Glama

Scrapling Fetch MCP

by cyberchitta
config.yaml2.16 kB
__info__: 'This project uses llm-context. For more information, visit: https://github.com/cyberchitta/llm-context.py or https://pypi.org/project/llm-context/' profiles: code: description: Default profile for software projects, selecting all code files while excluding media and git-related files. gitignores: full_files: - .git - .gitignore - .llm-context/ - '*.tmp' - '*.lock' - package-lock.json - yarn.lock - pnpm-lock.yaml - go.sum - elm-stuff - LICENSE - CHANGELOG.md - README.md - .env - .dockerignore - Dockerfile - docker-compose.yml - '*.log' - '*.svg' - '*.png' - '*.jpg' - '*.jpeg' - '*.gif' - '*.ico' - '*.woff' - '*.woff2' - '*.eot' - '*.ttf' - '*.map' outline_files: - .git - .gitignore - .llm-context/ - '*.tmp' - '*.lock' - package-lock.json - yarn.lock - pnpm-lock.yaml - go.sum - elm-stuff - LICENSE - CHANGELOG.md - README.md - .env - .dockerignore - Dockerfile - docker-compose.yml - '*.log' - '*.svg' - '*.png' - '*.jpg' - '*.jpeg' - '*.gif' - '*.ico' - '*.woff' - '*.woff2' - '*.eot' - '*.ttf' - '*.map' only-include: full_files: - '**/*' outline_files: - '**/*' prompt: lc-prompt.md settings: no_media: false with_prompt: false with_user_notes: false code-file: base: code description: Extends 'code' by saving the generated context to 'project-context.md.tmp' for external use. settings: context_file: project-context.md.tmp code-prompt: base: code description: Extends 'code' by including LLM instructions from lc-prompt.md for guided interactions. settings: with_prompt: true templates: context: lc-context.j2 context-mcp: lc-context-mcp.j2 definitions: lc-definitions.j2 files: lc-files.j2 highlights: lc-highlights.j2 prompt: lc-prompt.j2

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/cyberchitta/scrapling-fetch-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server