Provides community support and priority assistance through the DevDocs Discord server
Enables easy deployment and management of the DevDocs service using Docker containers, with scripts for startup, monitoring, and maintenance
Allows access to the DevDocs repository for contribution, with features for tracking star history and community engagement
DevDocs by CyberAGI 🚀
🚀 Technology Partners
🎯 Perfect For
🏢 Enterprise Software Developers
Skip weeks of reading documentation and dealing with technical debt. Implement ANY technology faster by letting DevDocs handle the heavy lifting of documentation understanding.
🕸️ Web Scrapers
Pull entire contents of websites with Smart Discovery of Child URLs up to level 5. Perfect for both internal and external website documentation with intelligent crawling.
👥 Development Teams
Leverage internal documentation with built-in MCP servers and Claude integration for intelligent data querying. Transform your team's knowledge base into an actionable resource.
🚀 Indie Hackers
DevDocs + VSCode(cline) + Your Idea = Ship products fast with ANY technology. No more getting stuck in documentation hell when building your next big thing.
✨ Features
🧠 Intelligent Crawling
Smart Depth Control: Choose crawl depth from 1-5 levels
Automatic Link Discovery: Finds and categorizes all related content
Selective Crawling: Pick exactly what you want to extract
Child URL Detection: Automatically discovers and maps website structure
⚡ Performance & Speed
Parallel Processing: Crawl multiple pages simultaneously
Smart Caching: Never waste time on duplicate content
Lazy Loading Support: Handles modern web apps effortlessly
Rate Limiting: Respectful crawling that won't overload servers
🎯 Content Processing
Clean Extraction: Get content without the fluff
Multiple Formats: Export to MD or JSON for LLM fine-tuning
Structured Output: Logically organized content
MCP Server Integration: Ready for AI processing
🛡️ Enterprise Features
Error Recovery: Auto-retry on failures
Full Logging: Track every operation
API Access: Integrate with your tools
Team Management: Multiple seats and roles
🤔 Why DevDocs?
The Problem
Documentation is everywhere and LLMs are OUTDATED in their knowledge. Reading it, understanding it, and implementing it takes weeks of research and development even for senior engineers. We cut down that time to hours.
Our Solution
DevDocs brings documentation to you. Point it at any tech documentation URL, and watch as it:
Discovers all related pages to that technology
Extracts meaningful content without the fluff
Organizes information logically inside an MCP server ready for your LLM to query
Presents it in a clean, searchable format in MD or JSON for finetuning LLM purpose
🔥 We want anyone in the world to have the ability to build amazing products quickly using the most cutting edge LLM technology.
💰 Pricing Comparison
Feature | DevDocs | Firecrawl |
Free Tier | Unlimited pages | None |
Starting Price | Free Forever | $16/month |
Enterprise Plan | Custom | $333/month |
Crawl Speed | 1000/min | 20/min |
Depth Levels | Up to 5 | Limited |
Team Seats | Unlimited | 1-5 seats |
Export Formats | MD, JSON, LLM-ready MCP servers | Limited formats |
API Access | Coming Soon | Limited |
Model Context Protocol Integration | ✅ | ❌ |
Support | Priority Available via Discord | Standard only |
Self-hosted (free use) | ✅ | ❌ |
🚀 Getting Started
DevDocs is designed to be easy to use with Docker, requiring minimal setup for new users.
Prerequisites
Docker installed on your system
Git for cloning the repository
Quick Start with Docker (Recommended)
For Mac/Linux users:
For Windows users: Experimental Only (Not Tested Yet)
If you encounter permission issues, you may need to run the script as administrator or manually set permissions on the logs, storage, and crawl_results directories. The script uses the
icacls
command to set permissions, which might require elevated privileges on some Windows systems.Manually Setting Permissions on Windows:
If you need to manually set permissions, you can do so using either the Windows GUI or command line:
Using Windows Explorer:
Right-click on each directory (logs, storage, crawl_results)
Select "Properties"
Go to the "Security" tab
Click "Edit" to change permissions
Click "Add" to add users/groups
Type "Everyone" and click "Check Names"
Click "OK"
Select "Everyone" in the list
Check "Full control" under "Allow"
Click "Apply" and "OK"
Using Command Prompt (as Administrator):
If you encounter issues with the docker-compose.yml file (such as "Top-level object must be a mapping" error), the
docker-start.bat
script automatically fixes this by ensuring the file has the correct format and encoding. This fix is applied every time you run the script, so you don't need to manually modify the file.
This single command will:
Create all necessary directories
Set appropriate permissions
Build and start all Docker containers
Monitor the services to ensure they're running properly
Accessing DevDocs
Once the services are running:
Frontend UI: http://localhost:3001
Backend API: http://localhost:24125
Crawl4AI Service: http://localhost:11235
Logs and Monitoring
When using Docker, logs can be accessed :
Container Logs (recommended for debugging):
To stop all services, press Ctrl+C
in the terminal where docker-start is running.
📜 Scripts and Their Purpose
DevDocs includes various utility scripts to help with development, testing, and maintenance. Here's a quick reference:
Startup Scripts
start.sh
/start.bat
/start.ps1
- Start all services (frontend, backend, MCP) for local development.docker-start.sh
/docker-start.bat
- Start all services using Docker containers.
MCP Server Scripts
check_mcp_health.sh
- Verify the MCP server's health and configuration status.restart_and_test_mcp.sh
- Restart Docker containers with updated MCP configuration and test connectivity.
Crawl4AI Scripts
check_crawl4ai.sh
- Check the status and health of the Crawl4AI service.debug_crawl4ai.sh
- Run Crawl4AI in debug mode with verbose logging for troubleshooting.test_crawl4ai.py
- Run tests against the Crawl4AI service to verify functionality.test_from_container.sh
- Test the Crawl4AI service from within a Docker container.
Utility Scripts
view_result.sh
- Display crawl results in a formatted view.find_empty_folders.sh
- Identify empty directories in the project structure.analyze_empty_folders.sh
- Analyze empty folders and categorize them by risk level.verify_reorganization.sh
- Verify that code reorganization was successful.
These scripts are organized in the following directories:
Root directory: Main scripts for common operations
scripts/general/
: General utility scriptsscripts/docker/
: Docker-specific scriptsscripts/mcp/
: MCP server management scriptsscripts/test/
: Testing and verification scripts
🌍 Built for Developers, by Developers
DevDocs is more than a tool—it's your documentation companion that:
Saves Time: Turn weeks of research into hours
Improves Understanding: Get clean, organized documentation
Enables Innovation: Build faster with any technology
Supports Teams: Share knowledge efficiently
LLM READY: Modern times require modern solutions, using devdocs with LLM is extremely easy and intuitive. With minimal configuration you can run Devdocs and Claude App and recognizes DevDocs's MCP server ready to chat with your data.
🛠️ Setting Up the Cline/Roo Cline for Rapid software development.
Open the "Modes" Interface
In Roo Code, click the + to create a new Mode-Specific Prompts.
Name
Give the mode a Name (e.g.,
Research_MCP
).
Role Definition Prompt
Mode-Specific Custom Instructions Prompt
🤝 Join Our Community
🏆 Success Stories
"DevDocs turned our 3-week implementation timeline into 2 days. It's not just a crawler, it's a development accelerator." - Senior Engineer at Fortune 100 Company
"Launched my SaaS in half the time by using DevDocs to understand and implement new technologies quickly." - Successful Indie Hacker
🛣️ DevDocs Roadmap
This roadmap outlines the upcoming enhancements and features planned for DevDocs, our advanced web crawling platform powered by Crawl4AI. Each item is designed to leverage Crawl4AI’s capabilities to their fullest, ensuring a robust, efficient, and user-friendly web crawling experience.
⸻
1. Enhanced Crawler Logic for Dynamic Content
Implement
wait_for_images=True
to ensure all images are fully loaded before extraction.Set
scan_full_page=True
to force the crawler to scroll through the entire page, triggering lazy-loaded content.Introduce
scroll_delay
to add delays between scroll steps, allowing content to load properly.Incorporate
wait_for
parameters to wait for specific DOM elements indicative of content loading completion.
2. Hot Loading with Browser Pooling
Implement a pool of pre-warmed browser instances to avoid the overhead of launching a new browser for each task.
Utilize
use_persistent_context=True
to maintain session data across tasks, reducing the need for repeated logins and setups.
3. Revamped Docker Containers with Latest DevDocs Integration
Update Docker images to incorporate the latest DevDocs features and optimizations.
Include environment variables for API tokens (
CRAWL4AI_API_TOKEN
) to secure API endpoints.Set appropriate memory limits and resource constraints to optimize performance.
4. Multi-OS Docker Instance Support
Create Docker images for different architectures (e.g.,
x86_64
,ARM
) to support a wide range of systems.Implement CI/CD pipelines to build and test images across multiple OS environments, ensuring compatibility and stability.
5. Memory-Adaptive Crawling
Integrate DevDocs’
MemoryAdaptiveDispatcher
to dynamically adjust concurrency based on system memory availability.Implement built-in rate limiting to prevent overwhelming target websites and avoid out-of-memory errors.
6. PDF Upload and Extraction in UI
Utilize DevDocs’ capability to export pages as PDFs (
pdf=True
) and extract content from them.Develop frontend components to handle PDF uploads, display extracted content, and allow users to interact with the data.
7. Hosted Environment with Persistent Storage and Enhanced UX
Implement BYO-databases solutions to store data privately, crawl results, and configurations across sessions.
Design intuitive dashboards and interfaces for users to manage their crawls, view results, and configure settings.
Ensure responsive design and accessibility across various browsers.
Star History
This server cannot be installed
hybrid server
The server is able to function both locally and remotely, depending on the configuration or use case.
Integrates with Claude to enable intelligent querying of documentation data, transforming crawled technical documentation into an actionable resource that LLMs can directly interact with.
Related MCP Servers
- -securityFlicense-qualityA smart documentation server that provides AI-assisted code improvement and documentation management through Claude Desktop integration.Last updated -10
Inkeep MCP Serverofficial
-securityAlicense-qualityA server that connects Claude to your documentation via Inkeep's API, enabling AI-powered interactions with your documentation content.Last updated -21MIT License- -securityFlicense-qualityAn MCP server that integrates with Claude to provide smart documentation search capabilities across multiple AI/ML libraries, allowing users to retrieve and process technical information through natural language queries.Last updated -
- AsecurityFlicenseAqualityA server that enables Claude to search and access documentation from popular libraries like LangChain, LlamaIndex, and OpenAI directly within conversations.Last updated -13