Compare commits

..

49 Commits

Author SHA1 Message Date
AHMET YILMAZ
eca04b0368 Refactor Pydantic model configuration to use ConfigDict for arbitrary types 2025-11-18 15:40:17 +08:00
ntohidi
c2c4d42be4 Fix #1181: Preserve whitespace in code blocks during HTML scraping
The remove_empty_elements_fast() method was removing whitespace-only
  span elements inside <pre> and <code> tags, causing import statements
  like "import torch" to become "importtorch". Now skips elements inside
  code blocks where whitespace is significant.
2025-11-17 12:21:23 +01:00
Aravind
f68e7531e3 Sponsors/scrapeless (#1619) 2025-11-17 07:44:52 +01:00
UncleCode
cb637fb5c4 Merge pull request #1613 from unclecode/release/v0.7.7 2025-11-16 12:26:54 +01:00
ntohidi
6244f56f36 Release v0.7.7
- Updated version to 0.7.7
- Added comprehensive demo and release notes
- Updated all documentation
2025-11-14 10:23:31 +01:00
ntohidi
2c973b1183 Merge branch 'develop' into release/v0.7.7 2025-11-13 14:54:05 +01:00
Nasrin
f3146de969 Merge pull request #1609 from unclecode/fix/update-config-documentation
Update browser and crawler run config documentation to match async_configs.py implementation
2025-11-13 21:52:53 +08:00
Soham Kukreti
d6b6d11a2d docs: update browser and crawler run config documentation to match async_configs.py implementation
Updated browser-crawler-config.md and parameters.md to ensure complete
accuracy with the actual BrowserConfig and CrawlerRunConfig implementations.

Changes:
- Removed non-existent parameters from documentation:
  * enable_rate_limiting, rate_limit_config (never implemented)
  * memory_threshold_percent, check_interval, max_session_permit (internal to AsyncDispatcher)
  * display_mode (doesn't exist)

- Added missing BrowserConfig parameters (14 total):
  * browser_mode, use_managed_browser, cdp_url, debugging_port, host
  * viewport, chrome_channel, channel
  * accept_downloads, downloads_path, storage_state, sleep_on_close
  * user_agent_mode, user_agent_generator_config, enable_stealth

- Added missing CrawlerRunConfig parameters (29 total):
  * chunking_strategy, keep_attrs, parser_type, scraping_strategy
  * proxy_config, proxy_rotation_strategy
  * locale, timezone_id, geolocation, fetch_ssl_certificate
  * shared_data, wait_for_timeout
  * c4a_script, max_scroll_steps
  * exclude_all_images, table_score_threshold, table_extraction
  * exclude_internal_links, score_links
  * capture_network_requests, capture_console_messages
  * method, stream, url, user_agent, user_agent_mode, user_agent_generator_config
  * deep_crawl_strategy, link_preview_config, url_matcher, match_mode, experimental

- Marked deprecated cache parameters (bypass_cache, disable_cache, no_cache_read, no_cache_write)
- Reorganized parameters into logical sections (Content Processing, Browser Location & Identity,
  Caching & Session, Page Navigation & Timing, Page Interaction, Media Handling, Link/Domain
  Handling, Debug & Logging, Connection & HTTP, Virtual Scroll, URL Matching, Advanced Features)
- Ensured all parameter descriptions match source code docstrings
- Added proper default values from __init__ signatures
2025-11-13 14:54:16 +05:30
ntohidi
b58579548c Bump version to 0.7.7 for stable release 2025-11-13 09:52:18 +01:00
Nasrin
466be69e72 Merge pull request #1607 from unclecode/fix/dfs_deep_crawling
Fix/dfs deep crawling
2025-11-13 16:43:47 +08:00
AHMET YILMAZ
ceade853c3 Enhance DFSDeepCrawlStrategy documentation for clarity and detail 2025-11-13 16:39:08 +08:00
ntohidi
998c809e08 Rename folder name for NSTProxy integration examples for crawl4ai 2025-11-13 09:36:39 +01:00
ntohidi
d0fb53540d Update proxy-security documentation 2025-11-13 09:23:44 +01:00
Nasrin
8116b15b63 Merge pull request #1596 from unclecode/docs-proxy-security
#1591 enhance proxy configuration with security, SSL analysis, and rotation examples
2025-11-13 16:22:28 +08:00
AHMET YILMAZ
fe353c4e27 Refactor proxy configuration documentation for clarity and consistency 2025-11-13 11:20:24 +08:00
ntohidi
89cc29fe44 Merge branch 'fix/docker' into develop 2025-11-12 17:06:31 +01:00
Nasrin
cdcb8836b7 Merge pull request #1605 from Nstproxy/feat/nstproxy
feat: Add Nstproxy Proxies
2025-11-12 23:56:14 +08:00
Nasrin
b207ae2848 Merge pull request #1528 from unclecode/fix/managed-browser-cdp-timing
Add CDP endpoint verification with exponential backoff for managed browsers
2025-11-12 23:53:57 +08:00
Nasrin
be00fc3a42 Merge pull request #1598 from unclecode/fix/sitemap_seeder
#1559 :Add tests for sitemap parsing and URL normalization in AsyncUr…
2025-11-12 18:09:34 +08:00
Nasrin
124ac583bb Merge pull request #1599 from unclecode/docs-llm-strategies-update
#1551 : Fix casing and variable name consistency for LLMConfig in doc…
2025-11-12 17:54:26 +08:00
AHMET YILMAZ
1bd3de6a47 #1510 : Add DFS deep crawler demonstration script and enhance DFS strategy with seen URL tracking 2025-11-12 17:44:43 +08:00
nstproxy
80452166c8 feat: Add Nstproxy Proxies 2025-11-12 16:25:39 +08:00
UncleCode
a99cd37c0e Merge pull request #1597 from unclecode/sponsors/capsolver 2025-11-11 14:50:44 +08:00
AHMET YILMAZ
2e8f8c9b49 #1551 : Fix casing and variable name consistency for LLMConfig in documentation 2025-11-10 15:38:14 +08:00
AHMET YILMAZ
80745bceb9 #1559 :Add tests for sitemap parsing and URL normalization in AsyncUrlSeeder 2025-11-10 14:15:54 +08:00
Aravind Karnam
4bee230c37 docs: Add a tip for captcha solving usecases using a third party integration 2025-11-10 11:20:48 +05:30
Aravind
006e29f308 Merge pull request #1589 from capsolver/main
Add some examples of using capsolver to solve captcha
2025-11-10 10:45:16 +05:30
AHMET YILMAZ
263ac890fd #1591
: Enhance proxy configuration documentation with security features, SSL analysis, and improved examples
2025-11-10 11:42:07 +08:00
unclecode
1a22fb4d4f docs: rename Docker deployment to self-hosting guide with comprehensive monitoring documentation
Major documentation restructuring to emphasize self-hosting capabilities and fully document the real-time monitoring system.

Changes:
- Renamed docker-deployment.md → self-hosting.md to better reflect the value proposition
- Updated mkdocs.yml navigation to "Self-Hosting Guide"
- Completely rewrote introduction emphasizing self-hosting benefits:
  * Data privacy and ownership
  * Cost control and transparency
  * Performance and security advantages
  * Full customization capabilities

- Expanded "Metrics & Monitoring" → "Real-time Monitoring & Operations" with:
  * Monitoring Dashboard section documenting the /monitor UI
  * Complete feature breakdown (system health, requests, browsers, janitor, errors)
  * Monitor API Endpoints with all REST endpoints and examples
  * WebSocket Streaming integration guide with Python examples
  * Control Actions for manual browser management
  * Production Integration patterns (Prometheus, custom dashboards, alerting)
  * Key production metrics to track

- Enhanced summary section:
  * What users learned checklist
  * Why self-hosting matters
  * Clear next steps
  * Key resources with monitoring dashboard URL

The monitoring dashboard built 2-3 weeks ago is now fully documented and discoverable.
Users will understand they have complete operational visibility at http://localhost:11235/monitor
with real-time updates, browser pool management, and programmatic control via REST/WebSocket APIs.

This positions Crawl4AI as an enterprise-grade self-hosting solution with DevOps-level
monitoring capabilities, not just a Docker deployment.
2025-11-09 13:31:52 +08:00
unclecode
81b5312629 Update gitignore 2025-11-09 10:49:42 +08:00
Nasrin
d56b0eb9a9 Merge pull request #1495 from unclecode/fix/viewport_in_managed_browser
feat(ManagedBrowser): add viewport size configuration for browser launch
2025-11-06 18:42:45 +08:00
Nasrin
66175e132b Merge pull request #1590 from unclecode/fix/async-llm-extraction-arunMany
This commit resolves issue #1055 where LLM extraction was blocking async
2025-11-06 18:40:42 +08:00
ntohidi
a30548a98f This commit resolves issue #1055 where LLM extraction was blocking async
execution, causing URLs to be processed sequentially instead of in parallel.

  Changes:
  - Added aperform_completion_with_backoff() using litellm.acompletion for async LLM calls
  - Implemented arun() method in ExtractionStrategy base class with thread pool fallback
  - Created async arun() and aextract() methods in LLMExtractionStrategy using asyncio.gather
  - Updated AsyncWebCrawler.arun() to detect and use arun() when available
  - Added comprehensive test suite to verify parallel execution

  Impact:
  - LLM extraction now runs truly in parallel across multiple URLs
  - Significant performance improvement for multi-URL crawls with LLM strategies
  - Backward compatible - existing extraction strategies continue to work
  - No breaking changes to public API

  Technical details:
  - Uses litellm.acompletion for non-blocking LLM calls
  - Leverages asyncio.gather for concurrent chunk processing
  - Maintains backward compatibility via asyncio.to_thread fallback
  - Works seamlessly with MemoryAdaptiveDispatcher and other dispatchers
2025-11-06 11:22:45 +01:00
CapSolver
2ae9899eac Clarify CapSolver integration instructions
Updated text for clarity and capitalization.
2025-11-06 15:49:30 +08:00
CapSolver
57aeb70f00 Add CapSolver Captcha Solver 2025-11-06 15:37:31 +08:00
Nasrin
2c918155aa Merge pull request #1529 from unclecode/fix/remove_overlay_elements
Fix remove_overlay_elements functionality by calling injected JS function.
2025-11-06 00:10:32 +08:00
Nasrin
854694ef33 Merge pull request #1537 from unclecode/fix/docker-compose-llm-env
fix(docker): Remove environment variable overrides in docker-compose.yml
2025-11-06 00:07:51 +08:00
Nasrin
6534ece026 Merge pull request #1532 from unclecode/fix/update-documentation
Standardize C4A-Script tutorial, add CLI identity-based crawling, and add sponsorship CTA
2025-11-05 23:37:05 +08:00
Nasrin
89e28d4eee Merge pull request #1558 from unclecode/claude/fix-update-pyopenssl-security-011CUPexU25DkNvoxfu5ZrnB
Claude/fix update pyopenssl security 011 cu pex u25 dk nvoxfu5 zrn b
2025-10-28 17:09:11 +08:00
ntohidi
c0f1865287 feat(api): update marketplace version and build date in root endpoint response 2025-10-26 11:35:39 +01:00
ntohidi
46ef1116c4 fix(app-detail): enhance tab functionality, hide documentation and support tabs in marketplace 2025-10-26 11:21:29 +01:00
Claude
613097d121 test: add verification tests for pyOpenSSL security update
- Add lightweight security test to verify version requirements
- Add comprehensive integration test for crawl4ai functionality
- Tests verify pyOpenSSL >= 25.3.0 and cryptography >= 45.0.7
- All tests passing: security vulnerability is resolved

Related to #1545

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-23 06:57:25 +00:00
Claude
44ef0682b0 fix: update pyOpenSSL to >=25.3.0 to address security vulnerability
- Updates pyOpenSSL from >=24.3.0 to >=25.3.0
- This resolves CVE affecting cryptography package versions >=37.0.0 & <43.0.1
- pyOpenSSL 25.3.0 requires cryptography>=45.0.7, which is above the vulnerable range
- Fixes issue #1545

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-23 06:51:25 +00:00
Nasrin
40173eeb73 Update Docker hooks and Webhook documents (#1557)
* fix(docker-api): migrate to modern datetime library API

Signed-off-by: Emmanuel Ferdman <emmanuelferdman@gmail.com>

* Fix examples in README.md

* feat(docker): add user-provided hooks support to Docker API

Implements comprehensive hooks functionality allowing users to provide custom Python
functions as strings that execute at specific points in the crawling pipeline.

Key Features:
- Support for all 8 crawl4ai hook points:
  • on_browser_created: Initialize browser settings
  • on_page_context_created: Configure page context
  • before_goto: Pre-navigation setup
  • after_goto: Post-navigation processing
  • on_user_agent_updated: User agent modification handling
  • on_execution_started: Crawl execution initialization
  • before_retrieve_html: Pre-extraction processing
  • before_return_html: Final HTML processing

Implementation Details:
- Created UserHookManager for validation, compilation, and safe execution
- Added IsolatedHookWrapper for error isolation and timeout protection
- AST-based validation ensures code structure correctness
- Sandboxed execution with restricted builtins for security
- Configurable timeout (1-120 seconds) prevents infinite loops
- Comprehensive error handling ensures hooks don't crash main process
- Execution tracking with detailed statistics and logging

API Changes:
- Added HookConfig schema with code and timeout fields
- Extended CrawlRequest with optional hooks parameter
- Added /hooks/info endpoint for hook discovery
- Updated /crawl and /crawl/stream endpoints to support hooks

Safety Features:
- Malformed hooks return clear validation errors
- Hook errors are isolated and reported without stopping crawl
- Execution statistics track success/failure/timeout rates
- All hook results are JSON-serializable

Testing:
- Comprehensive test suite covering all 8 hooks
- Error handling and timeout scenarios validated
- Authentication, performance, and content extraction examples
- 100% success rate in production testing

Documentation:
- Added extensive hooks section to docker-deployment.md
- Security warnings about user-provided code risks
- Real-world examples using httpbin.org, GitHub, BBC
- Best practices and troubleshooting guide

ref #1377

* fix(deep-crawl): BestFirst priority inversion; remove pre-scoring truncation. ref #1253

  Use negative scores in PQ to visit high-score URLs first and drop link cap prior to scoring; add test for ordering.

* docs: Update URL seeding examples to use proper async context managers
- Wrap all AsyncUrlSeeder usage with async context managers
- Update URL seeding adventure example to use "sitemap+cc" source, focus on course posts, and add stream=True parameter to fix runtime error

* fix(crawler): Removed the incorrect reference in browser_config variable #1310

* docs: update Docker instructions to use the latest release tag

* fix(docker): Fix LLM API key handling for multi-provider support

Previously, the system incorrectly used OPENAI_API_KEY for all LLM providers
due to a hardcoded api_key_env fallback in config.yml. This caused authentication
errors when using non-OpenAI providers like Gemini.

Changes:
- Remove api_key_env from config.yml to let litellm handle provider-specific env vars
- Simplify get_llm_api_key() to return None, allowing litellm to auto-detect keys
- Update validate_llm_provider() to trust litellm's built-in key detection
- Update documentation to reflect the new automatic key handling

The fix leverages litellm's existing capability to automatically find the correct
environment variable for each provider (OPENAI_API_KEY, GEMINI_API_TOKEN, etc.)
without manual configuration.

ref #1291

* docs: update adaptive crawler docs and cache defaults; remove deprecated examples (#1330)
- Replace BaseStrategy with CrawlStrategy in custom strategy examples (DomainSpecificStrategy, HybridStrategy)
- Remove “Custom Link Scoring” and “Caching Strategy” sections no longer aligned with current library
- Revise memory pruning example to use adaptive.get_relevant_content and index-based retention of top 500 docs
- Correct Quickstart note: default cache mode is CacheMode.BYPASS; instruct enabling with CacheMode.ENABLED

* fix(utils): Improve URL normalization by avoiding quote/unquote to preserve '+' signs. ref #1332

* feat: Add comprehensive website to API example with frontend

This commit adds a complete, web scraping API example that demonstrates how to get structured data from any website and use it like an API using the crawl4ai library with a minimalist frontend interface.

Core Functionality
- AI-powered web scraping with plain English queries
- Dual scraping approaches: Schema-based (faster) and LLM-based (flexible)
- Intelligent schema caching for improved performance
- Custom LLM model support with API key management
- Automatic duplicate request prevention

Modern Frontend Interface
- Minimalist black-and-white design inspired by modern web apps
- Responsive layout with smooth animations and transitions
- Three main pages: Scrape Data, Models Management, API Request History
- Real-time results display with JSON formatting
- Copy-to-clipboard functionality for extracted data
- Toast notifications for user feedback
- Auto-scroll to results when scraping starts

Model Management System
- Web-based model configuration interface
- Support for any LLM provider (OpenAI, Gemini, Anthropic, etc.)
- Simplified configuration requiring only provider and API token
- Add, list, and delete model configurations
- Secure storage of API keys in local JSON files

API Request History
- Automatic saving of all API requests and responses
- Display of request history with URL, query, and cURL commands
- Duplicate prevention (same URL + query combinations)
- Request deletion functionality
- Clean, simplified display focusing on essential information

Technical Implementation

Backend (FastAPI)
- RESTful API with comprehensive endpoints
- Pydantic models for request/response validation
- Async web scraping with crawl4ai library
- Error handling with detailed error messages
- File-based storage for models and request history

Frontend (Vanilla JS/CSS/HTML)
- No framework dependencies - pure HTML, CSS, JavaScript
- Modern CSS Grid and Flexbox layouts
- Custom dropdown styling with SVG arrows
- Responsive design for mobile and desktop
- Smooth scrolling and animations

Core Library Integration
- WebScraperAgent class for orchestration
- ModelConfig class for LLM configuration management
- Schema generation and caching system
- LLM extraction strategy support
- Browser configuration with headless mode

* fix(dependencies): add cssselect to project dependencies

Fixes bug reported in issue #1405
[Bug]: Excluded selector (excluded_selector) doesn't work

This commit reintroduces the cssselect library which was removed by PR (https://github.com/unclecode/crawl4ai/pull/1368) and merged via (437395e490).

Integration tested against 0.7.4 Docker container. Reintroducing cssselector package eliminated errors seen in logs and excluded_selector functionality was restored.

Refs: #1405

* fix(docker): resolve filter serialization and JSON encoding errors in deep crawl strategy (ref #1419)

  - Fix URLPatternFilter serialization by preventing private __slots__ from being serialized as constructor params
  - Add public attributes to URLPatternFilter to store original constructor parameters for proper serialization
  - Handle property descriptors in CrawlResult.model_dump() to prevent JSON serialization errors
  - Ensure filter chains work correctly with Docker client and REST API

  The issue occurred because:
  1. Private implementation details (_simple_suffixes, etc.) were being serialized and passed as constructor arguments during deserialization
  2. Property descriptors were being included in the serialized output, causing "Object of type property is not JSON serializable" errors

  Changes:
  - async_configs.py: Comment out __slots__ serialization logic (lines 100-109)
  - filters.py: Add patterns, use_glob, reverse to URLPatternFilter __slots__ and store as public attributes
  - models.py: Convert property descriptors to strings in model_dump() instead of including them directly

* fix(logger): ensure logger is a Logger instance in crawling strategies. ref #1437

* feat(docker): Add temperature and base_url parameters for LLM configuration. ref #1035

  Implement hierarchical configuration for LLM parameters with support for:
  - Temperature control (0.0-2.0) to adjust response creativity
  - Custom base_url for proxy servers and alternative endpoints
  - 4-tier priority: request params > provider env > global env > defaults

  Add helper functions in utils.py, update API schemas and handlers,
  support environment variables (LLM_TEMPERATURE, OPENAI_TEMPERATURE, etc.),
  and provide comprehensive documentation with examples.

* feat(docker): improve docker error handling
- Return comprehensive error messages along with status codes for api internal errors.
- Fix fit_html property serialization issue in both /crawl and /crawl/stream endpoints
- Add sanitization to ensure fit_html is always JSON-serializable (string or None)
- Add comprehensive error handling test suite.

* #1375 : refactor(proxy) Deprecate 'proxy' parameter in BrowserConfig and enhance proxy string parsing

- Updated ProxyConfig.from_string to support multiple proxy formats, including URLs with credentials.
- Deprecated the 'proxy' parameter in BrowserConfig, replacing it with 'proxy_config' for better flexibility.
- Added warnings for deprecated usage and clarified behavior when both parameters are provided.
- Updated documentation and tests to reflect changes in proxy configuration handling.

* Remove deprecated test for 'proxy' parameter in BrowserConfig and update .gitignore to include test_scripts directory.

* feat: add preserve_https_for_internal_links flag to maintain HTTPS during crawling. Ref #1410

Added a new `preserve_https_for_internal_links` configuration flag that preserves the original HTTPS scheme for same-domain links even when the server redirects to HTTP.

* feat: update documentation for preserve_https_for_internal_links. ref #1410

* fix: drop Python 3.9 support and require Python >=3.10.
The library no longer supports Python 3.9 and so it was important to drop all references to python 3.9.
Following changes have been made:
- pyproject.toml: set requires-python to ">=3.10"; remove 3.9 classifier
- setup.py: set python_requires to ">=3.10"; remove 3.9 classifier
- docs: update Python version mentions
  - deploy/docker/c4ai-doc-context.md: options -> 3.10, 3.11, 3.12, 3.13

* issue #1329 refactor(crawler): move unwanted properties to CrawlerRunConfig class

* fix(auth): fixed Docker JWT authentication. ref #1442

* remove: delete unused yoyo snapshot subproject

* fix: raise error on last attempt failure in perform_completion_with_backoff. ref #989

* Commit without API

* fix: update option labels in request builder for clarity

* fix: allow custom LLM providers for adaptive crawler embedding config. ref: #1291

  - Change embedding_llm_config from Dict to Union[LLMConfig, Dict] for type safety
  - Add backward-compatible conversion property _embedding_llm_config_dict
  - Replace all hardcoded OpenAI embedding configs with configurable options
  - Fix LLMConfig object attribute access in query expansion logic
  - Add comprehensive example demonstrating multiple provider configurations
  - Update documentation with both LLMConfig object and dictionary usage patterns

  Users can now specify any LLM provider for query expansion in embedding strategy:
  - New: embedding_llm_config=LLMConfig(provider='anthropic/claude-3', api_token='key')
  - Old: embedding_llm_config={'provider': 'openai/gpt-4', 'api_token': 'key'} (still works)

* refactor(BrowserConfig): change deprecation warning for 'proxy' parameter to UserWarning

* feat(StealthAdapter): fix stealth features for Playwright integration. ref #1481

* #1505 fix(api): update config handling to only set base config if not provided by user

* fix(docker-deployment): replace console.log with print for metadata extraction

* Release v0.7.5: The Update

- Updated version to 0.7.5
- Added comprehensive demo and release notes
- Updated documentation

* refactor(release): remove memory management section for cleaner documentation. ref #1443

* feat(docs): add brand book and page copy functionality

- Add comprehensive brand book with color system, typography, components
- Add page copy dropdown with markdown copy/view functionality
- Update mkdocs.yml with new assets and branding navigation
- Use terminal-style ASCII icons and condensed menu design

* Update gitignore add local scripts folder

* fix: remove this import as it causes python to treat "json" as a variable in the except block

* fix: always return a list, even if we catch an exception

* feat(marketplace): Add Crawl4AI marketplace with secure configuration

- Implement marketplace frontend and admin dashboard
- Add FastAPI backend with environment-based configuration
- Use .env file for secrets management
- Include data generation scripts
- Add proper CORS configuration
- Remove hardcoded password from admin login
- Update gitignore for security

* fix(marketplace): Update URLs to use /marketplace path and relative API endpoints

- Change API_BASE to relative '/api' for production
- Move marketplace to /marketplace instead of /marketplace/frontend
- Update MkDocs navigation
- Fix logo path in marketplace index

* fix(docs): hide copy menu on non-markdown pages

* feat(marketplace): add sponsor logo uploads

Co-authored-by: factory-droid[bot] <138933559+factory-droid[bot]@users.noreply.github.com>

* feat(docs): add chatgpt quick link to page actions

* fix(marketplace): align admin api with backend endpoints

* fix(marketplace): isolate api under marketplace prefix

* fix(marketplace): resolve app detail page routing and styling issues

- Fixed JavaScript errors from missing HTML elements (install-code, usage-code, integration-code)
- Added missing CSS classes for tabs, overview layout, sidebar, and integration content
- Fixed tab navigation to display horizontally in single line
- Added proper padding to tab content sections (removed from container, added to content)
- Fixed tab selector from .nav-tab to .tab-btn to match HTML structure
- Added sidebar styling with stats grid and metadata display
- Improved responsive design with mobile-friendly tab scrolling
- Fixed code block positioning for copy buttons
- Removed margin from first headings to prevent extra spacing
- Added null checks for DOM elements in JavaScript to prevent errors

These changes resolve the routing issue where clicking on apps caused page redirects,
and fix the broken layout where CSS was not properly applied to the app detail page.

* fix(marketplace): prevent hero image overflow and secondary card stretching

- Fixed hero image to 200px height with min/max constraints
- Added object-fit: cover to hero-image img elements
- Changed secondary-featured align-items from stretch to flex-start
- Fixed secondary-card height to 118px (no flex: 1 stretching)
- Updated responsive grid layouts for wider screens
- Added flex: 1 to hero-content for better content distribution

These changes ensure a rigid, predictable layout that prevents:
1. Large images from pushing text content down
2. Single secondary cards from stretching to fill entire height

* feat: Add hooks utility for function-based hooks with Docker client integration. ref #1377

   Add hooks_to_string() utility function that converts Python function objects
   to string representations for the Docker API, enabling developers to write hooks
   as regular Python functions instead of strings.

   Core Changes:
   - New hooks_to_string() utility in crawl4ai/utils.py using inspect.getsource()
   - Docker client now accepts both function objects and strings for hooks
   - Automatic detection and conversion in Crawl4aiDockerClient._prepare_request()
   - New hooks and hooks_timeout parameters in client.crawl() method

   Documentation:
   - Docker client examples with function-based hooks (docs/examples/docker_client_hooks_example.py)
   - Updated main Docker deployment guide with comprehensive hooks section
   - Added unit tests for hooks utility (tests/docker/test_hooks_utility.py)

* feat: Add hooks utility for function-based hooks with Docker client integration. ref #1377

   Add hooks_to_string() utility function that converts Python function objects
   to string representations for the Docker API, enabling developers to write hooks
   as regular Python functions instead of strings.

   Core Changes:
   - New hooks_to_string() utility in crawl4ai/utils.py using inspect.getsource()
   - Docker client now accepts both function objects and strings for hooks
   - Automatic detection and conversion in Crawl4aiDockerClient._prepare_request()
   - New hooks and hooks_timeout parameters in client.crawl() method

   Documentation:
   - Docker client examples with function-based hooks (docs/examples/docker_client_hooks_example.py)
   - Updated main Docker deployment guide with comprehensive hooks section
   - Added unit tests for hooks utility (tests/docker/test_hooks_utility.py)

* fix(docs): clarify Docker Hooks System with function-based API in README

* docs: Add demonstration files for v0.7.5 release, showcasing the new Docker Hooks System and all other features.

* docs: Update 0.7.5 video walkthrough

* docs: add complete SDK reference documentation

Add comprehensive single-page SDK reference combining:
- Installation & Setup
- Quick Start
- Core API (AsyncWebCrawler, arun, arun_many, CrawlResult)
- Configuration (BrowserConfig, CrawlerConfig, Parameters)
- Crawling Patterns
- Content Processing (Markdown, Fit Markdown, Selection, Interaction, Link & Media)
- Extraction Strategies (LLM and No-LLM)
- Advanced Features (Session Management, Hooks & Auth)

Generated using scripts/generate_sdk_docs.py in ultra-dense mode
optimized for AI assistant consumption.

Stats: 23K words, 185 code blocks, 220KB

* feat: add AI assistant skill package for Crawl4AI

- Create comprehensive skill package for AI coding assistants
- Include complete SDK reference (23K words, v0.7.4)
- Add three extraction scripts (basic, batch, pipeline)
- Implement version tracking in skill and scripts
- Add prominent download section on homepage
- Place skill in docs/assets for web distribution

The skill enables AI assistants like Claude, Cursor, and Windsurf
to effectively use Crawl4AI with optimized workflows for markdown
generation and data extraction.

* fix: remove non-existent wiki link and clarify skill usage instructions

* fix: update Crawl4AI skill with corrected parameters and examples

- Fixed CrawlerConfig → CrawlerRunConfig throughout
- Fixed parameter names (timeout → page_timeout, store_html removed)
- Fixed schema format (selector → baseSelector)
- Corrected proxy configuration (in BrowserConfig, not CrawlerRunConfig)
- Fixed fit_markdown usage with content filters
- Added comprehensive references to docs/examples/ directory
- Created safe packaging script to avoid root directory pollution
- All scripts tested and verified working

* fix: thoroughly verify and fix all Crawl4AI skill examples

- Cross-checked every section against actual docs
- Fixed BM25ContentFilter parameters (user_query, bm25_threshold)
- Removed incorrect wait_for selector from basic example
- Added comprehensive test suite (4 test files)
- All examples now tested and verified working
- Tests validate: basic crawling, markdown generation, data extraction, advanced patterns
- Package size: 76.6 KB (includes tests for future validation)

* feat(ci): split release pipeline and add Docker caching

- Split release.yml into PyPI/GitHub release and Docker workflows
- Add GitHub Actions cache for Docker builds (10-15x faster rebuilds)
- Implement dual-trigger for docker-release.yml (auto + manual)
- Add comprehensive workflow documentation in .github/workflows/docs/
- Backup original workflow as release.yml.backup

* feat: add webhook notifications for crawl job completion

Implements webhook support for the crawl job API to eliminate polling requirements.

Changes:
- Added WebhookConfig and WebhookPayload schemas to schemas.py
- Created webhook.py with WebhookDeliveryService class
- Integrated webhook notifications in api.py handle_crawl_job
- Updated job.py CrawlJobPayload to accept webhook_config
- Added webhook configuration section to config.yml
- Included comprehensive usage examples in WEBHOOK_EXAMPLES.md

Features:
- Webhook notifications on job completion (success/failure)
- Configurable data inclusion in webhook payload
- Custom webhook headers support
- Global default webhook URL configuration
- Exponential backoff retry logic (5 attempts: 1s, 2s, 4s, 8s, 16s)
- 30-second timeout per webhook call

Usage:
POST /crawl/job with optional webhook_config:
- webhook_url: URL to receive notifications
- webhook_data_in_payload: include full results (default: false)
- webhook_headers: custom headers for authentication

Generated with Claude Code https://claude.com/claude-code

Co-Authored-By: Claude <noreply@anthropic.com>

* docs: add webhook documentation to Docker README

Added comprehensive webhook section to README.md including:
- Overview of asynchronous job queue with webhooks
- Benefits and use cases
- Quick start examples
- Webhook authentication
- Global webhook configuration
- Job status polling alternative

Updated table of contents and summary to include webhook feature.
Maintains consistent tone and style with rest of README.

Generated with Claude Code https://claude.com/claude-code

Co-Authored-By: Claude <noreply@anthropic.com>

* docs: add webhook example for Docker deployment

Added docker_webhook_example.py demonstrating:
- Submitting crawl jobs with webhook configuration
- Flask-based webhook receiver implementation
- Three usage patterns:
  1. Webhook notification only (fetch data separately)
  2. Webhook with full data in payload
  3. Traditional polling approach for comparison

Includes comprehensive comments explaining:
- Webhook payload structure
- Authentication headers setup
- Error handling
- Production deployment tips

Example is fully functional and ready to run with Flask installed.

Generated with Claude Code https://claude.com/claude-code

Co-Authored-By: Claude <noreply@anthropic.com>

* test: add webhook implementation validation tests

Added comprehensive test suite to validate webhook implementation:
- Module import verification
- WebhookDeliveryService initialization
- Pydantic model validation (WebhookConfig)
- Payload construction logic
- Exponential backoff calculation
- API integration checks

All tests pass (6/6), confirming implementation is correct.

Generated with Claude Code https://claude.com/claude-code

Co-Authored-By: Claude <noreply@anthropic.com>

* test: add comprehensive webhook feature test script

Added end-to-end test script that automates webhook feature testing:

Script Features (test_webhook_feature.sh):
- Automatic branch switching and dependency installation
- Redis and server startup/shutdown management
- Webhook receiver implementation
- Integration test for webhook notifications
- Comprehensive cleanup and error handling
- Returns to original branch after completion

Test Flow:
1. Fetch and checkout webhook feature branch
2. Activate venv and install dependencies
3. Start Redis and Crawl4AI server
4. Submit crawl job with webhook config
5. Verify webhook delivery and payload
6. Clean up all processes and return to original branch

Documentation:
- WEBHOOK_TEST_README.md with usage instructions
- Troubleshooting guide
- Exit codes and safety features

Usage: ./tests/test_webhook_feature.sh

Generated with Claude Code https://claude.com/claude-code

Co-Authored-By: Claude <noreply@anthropic.com>

* fix: properly serialize Pydantic HttpUrl in webhook config

Use model_dump(mode='json') instead of deprecated dict() method to ensure
Pydantic special types (HttpUrl, UUID, etc.) are properly serialized to
JSON-compatible native Python types.

This fixes webhook delivery failures caused by HttpUrl objects remaining
as Pydantic types in the webhook_config dict, which caused JSON
serialization errors and httpx request failures.

Also update mcp requirement to >=1.18.0 for compatibility.

* feat: add webhook support for /llm/job endpoint

Add comprehensive webhook notification support for the /llm/job endpoint,
following the same pattern as the existing /crawl/job implementation.

Changes:
- Add webhook_config field to LlmJobPayload model (job.py)
- Implement webhook notifications in process_llm_extraction() with 4
  notification points: success, provider validation failure, extraction
  failure, and general exceptions (api.py)
- Store webhook_config in Redis task data for job tracking
- Initialize WebhookDeliveryService with exponential backoff retry logic
Documentation:
- Add Example 6 to WEBHOOK_EXAMPLES.md showing LLM extraction with webhooks
- Update Flask webhook handler to support both crawl and llm_extraction tasks
- Add TypeScript client examples for LLM jobs
- Add comprehensive examples to docker_webhook_example.py with schema support
- Clarify data structure differences between webhook and API responses

Testing:
- Add test_llm_webhook_feature.py with 7 validation tests (all passing)
- Verify pattern consistency with /crawl/job implementation
- Add implementation guide (WEBHOOK_LLM_JOB_IMPLEMENTATION.md)

* fix: remove duplicate comma in webhook_config parameter

* fix: update Crawl4AI Docker container port from 11234 to 11235

* docs: enhance README and docker-deployment documentation with Job Queue and Webhook API details

* docs: update docker_hooks_examples.py with comprehensive examples and improved structure

---------

Signed-off-by: Emmanuel Ferdman <emmanuelferdman@gmail.com>
Co-authored-by: Emmanuel Ferdman <emmanuelferdman@gmail.com>
Co-authored-by: Nezar Ali <abu5sohaib@gmail.com>
Co-authored-by: Soham Kukreti <kukretisoham@gmail.com>
Co-authored-by: James T. Wood <jamesthomaswood@gmail.com>
Co-authored-by: AHMET YILMAZ <tawfik@kidocode.com>
Co-authored-by: nafeqq-1306 <nafiquee@yahoo.com>
Co-authored-by: unclecode <unclecode@kidocode.com>
Co-authored-by: Martin Sjöborg <martin.sjoborg@quartr.se>
Co-authored-by: Martin Sjöborg <martin@sjoborg.org>
Co-authored-by: factory-droid[bot] <138933559+factory-droid[bot]@users.noreply.github.com>
Co-authored-by: Claude <noreply@anthropic.com>
2025-10-22 22:34:19 +08:00
Soham Kukreti
46e1a67f61 fix(docker): Remove environment variable overrides in docker-compose.yml (#1411)
The docker-compose.yml had an `environment:` section with variable
substitutions (${VAR:-}) that was overriding values from .llm.env with
empty strings.

- Commented out the `environment:` section to prevent overwrites
- Added clear warning comment explaining the override behavior
- .llm.env values now load directly into container without interference
2025-10-06 14:41:22 +05:30
Soham Kukreti
7dfe528d43 fix(docs): standardize C4A-Script tutorial, add CLI identity-based crawling, and add sponsorship CTA
- Switch installs to pip install -r requirements.txt (tutorial and app docs)
- Update local run steps to python server.py and http://localhost:8000
- Set default PORT to 8000; update port-in-use commands and alt port 8001
- Replace unsupported :contains() example with accessible attribute selector
- Update example URLs in tutorial servers to 127.0.0.1:8000
- Add “Identity-based crawling” section with crwl profiles CLI workflow and code usage
- Replace legacy-docs note with sponsorship message in docs/md_v2/index.md
- Minor copy and consistency fixes across pages
2025-10-03 22:00:46 +05:30
Soham Kukreti
2dc6588573 fix: remove_overlay_elements functionality by calling injected JS function. ref: #1396
- Fix critical bug where overlay removal JS function was injected but never called
  - Change remove_overlay_elements() to properly execute the injected async function
  - Wrap JS execution in async to handle the async overlay removal logic
  - Add test_remove_overlay_elements() test case to verify functionality works
  - Ensure overlay elements (cookie banners, popups, modals) are actually removed

  The remove_overlay_elements feature now works as intended:
  - Before: Function definition injected but never executed (silent failure)
  - After: Function injected and called, successfully removing overlay elements
2025-09-29 20:40:08 +05:30
Soham Kukreti
34c0996ee4 fix: Add CDP endpoint verification with exponential backoff for managed browsers (#1445)
browser_manager:
- Add CDP endpoint verification with retry logic and exponential backoff
- Call verification before connecting to CDP in `start()` method
- Graceful handling of timing issues during browser startup

test_cdp_strategy:
- Fix cookie persistence test by adding storage state management
- Fix session management test to work with managed browser architecture
- Add comprehensive CDP timing tests covering:
  - Fast startup scenarios
  - Delayed browser startup simulation
  - Exponential backoff behavior validation
  - Concurrent browser connections
  - Stress testing with multiple successive startups
  - Retry count verification

Impact:
- Eliminates browser startup failures due to CDP timing issues
- Provides robust fallback with automatic retries
- Maintains fast startup when CDP is immediately available
- Comprehensive test coverage ensures reliability

Resolves CDP connection timing issues in managed browser mode.
2025-09-29 19:31:09 +05:30
AHMET YILMAZ
e3467c08f6 #1490 feat(ManagedBrowser): add viewport size configuration for browser launch 2025-09-17 17:40:38 +08:00
141 changed files with 6536 additions and 15319 deletions

View File

@@ -1,31 +0,0 @@
#!/bin/bash
# Pre-commit hook: Auto-sync cnode files when cnode source is modified
# Colors
RED='\033[0;31m'
GREEN='\033[0;32m'
YELLOW='\033[1;33m'
NC='\033[0m'
# Check if cnode source files are being committed
CNODE_FILES_CHANGED=$(git diff --cached --name-only | grep -E "deploy/docker/(cnode_cli|server_manager)\.py")
if [ -n "$CNODE_FILES_CHANGED" ]; then
echo -e "${YELLOW}🔄 cnode source files modified, auto-syncing to package...${NC}"
# Run sync script
if [ -f "deploy/installer/sync-cnode.sh" ]; then
bash deploy/installer/sync-cnode.sh
# Stage the synced files
git add deploy/installer/cnode_pkg/cli.py
git add deploy/installer/cnode_pkg/server_manager.py
echo -e "${GREEN}✅ cnode package synced and staged${NC}"
else
echo -e "${RED}❌ Error: sync-cnode.sh not found${NC}"
exit 1
fi
fi
exit 0

6
.gitignore vendored
View File

@@ -185,8 +185,7 @@ Crawl4AI.egg-info/
requirements0.txt
a.txt
# Ignore shell scripts globally, but allow test scripts
# *.sh
*.sh
.idea
docs/examples/.chainlit/
docs/examples/.chainlit/*
@@ -272,6 +271,8 @@ continue_config.json
CLAUDE_MONITOR.md
CLAUDE.md
.claude/
tests/**/test_site
tests/**/reports
tests/**/benchmark_reports
@@ -294,4 +295,3 @@ scripts/
*.db
*.rdb
*.ldb
.context/

View File

@@ -1,7 +1,7 @@
FROM python:3.12-slim-bookworm AS build
# C4ai version
ARG C4AI_VER=0.7.6
ARG C4AI_VER=0.7.7
ENV C4AI_VERSION=$C4AI_VER
LABEL c4ai.version=$C4AI_VER

View File

@@ -27,13 +27,13 @@
Crawl4AI turns the web into clean, LLM ready Markdown for RAG, agents, and data pipelines. Fast, controllable, battle tested by a 50k+ star community.
[✨ Check out latest update v0.7.6](#-recent-updates)
[✨ Check out latest update v0.7.7](#-recent-updates)
**New in v0.7.6**: Complete Webhook Infrastructure for Docker Job Queue API! Real-time notifications for both `/crawl/job` and `/llm/job` endpoints with exponential backoff retry, custom headers, and flexible delivery modes. No more polling! [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.6.md)
**New in v0.7.7**: Complete Self-Hosting Platform with Real-time Monitoring! Enterprise-grade monitoring dashboard, comprehensive REST API, WebSocket streaming, smart browser pool management, and production-ready observability. Full visibility and control over your crawling infrastructure. [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.7.md)
✨ Recent v0.7.5: Docker Hooks System with function-based API for pipeline customization, Enhanced LLM Integration with custom providers, HTTPS Preservation, and multiple community-reported bug fixes. [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.5.md)
✨ Recent v0.7.6: Complete Webhook Infrastructure for Docker Job Queue API! Real-time notifications for both `/crawl/job` and `/llm/job` endpoints with exponential backoff retry, custom headers, and flexible delivery modes. No more polling! [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.6.md)
✨ Previous v0.7.4: Revolutionary LLM Table Extraction with intelligent chunking, enhanced concurrency fixes, memory management refactor, and critical stability improvements. [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.4.md)
✨ Previous v0.7.5: Docker Hooks System with function-based API for pipeline customization, Enhanced LLM Integration with custom providers, HTTPS Preservation, and multiple community-reported bug fixes. [Release notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.5.md)
<details>
<summary>🤓 <strong>My Personal Story</strong></summary>
@@ -296,6 +296,7 @@ pip install -e ".[all]" # Install all optional features
### New Docker Features
The new Docker implementation includes:
- **Real-time Monitoring Dashboard** with live system metrics and browser pool visibility
- **Browser pooling** with page pre-warming for faster response times
- **Interactive playground** to test and generate request code
- **MCP integration** for direct connection to AI tools like Claude Code
@@ -310,7 +311,8 @@ The new Docker implementation includes:
docker pull unclecode/crawl4ai:latest
docker run -d -p 11235:11235 --name crawl4ai --shm-size=1g unclecode/crawl4ai:latest
# Visit the playground at http://localhost:11235/playground
# Visit the monitoring dashboard at http://localhost:11235/dashboard
# Or the playground at http://localhost:11235/playground
```
### Quick Test
@@ -339,7 +341,7 @@ else:
result = requests.get(f"http://localhost:11235/task/{task_id}")
```
For more examples, see our [Docker Examples](https://github.com/unclecode/crawl4ai/blob/main/docs/examples/docker_example.py). For advanced configuration, environment variables, and usage examples, see our [Docker Deployment Guide](https://docs.crawl4ai.com/basic/docker-deployment/).
For more examples, see our [Docker Examples](https://github.com/unclecode/crawl4ai/blob/main/docs/examples/docker_example.py). For advanced configuration, monitoring features, and production deployment, see our [Self-Hosting Guide](https://docs.crawl4ai.com/core/self-hosting/).
</details>
@@ -544,8 +546,63 @@ async def test_news_crawl():
</details>
---
> **💡 Tip:** Some websites may use **CAPTCHA** based verification mechanisms to prevent automated access. If your workflow encounters such challenges, you may optionally integrate a third-party CAPTCHA-handling service such as <strong>[CapSolver](https://www.capsolver.com/blog/Partners/crawl4ai-capsolver/?utm_source=crawl4ai&utm_medium=github_pr&utm_campaign=crawl4ai_integration)</strong>. They support reCAPTCHA v2/v3, Cloudflare Turnstile, Challenge, AWS WAF, and more. Please ensure that your usage complies with the target websites terms of service and applicable laws.
## ✨ Recent Updates
<details>
<summary><strong>Version 0.7.7 Release Highlights - The Self-Hosting & Monitoring Update</strong></summary>
- **📊 Real-time Monitoring Dashboard**: Interactive web UI with live system metrics and browser pool visibility
```python
# Access the monitoring dashboard
# Visit: http://localhost:11235/dashboard
# Real-time metrics include:
# - System health (CPU, memory, network, uptime)
# - Active and completed request tracking
# - Browser pool management (permanent/hot/cold)
# - Janitor cleanup events
# - Error monitoring with full context
```
- **🔌 Comprehensive Monitor API**: Complete REST API for programmatic access to all monitoring data
```python
import httpx
async with httpx.AsyncClient() as client:
# System health
health = await client.get("http://localhost:11235/monitor/health")
# Request tracking
requests = await client.get("http://localhost:11235/monitor/requests")
# Browser pool status
browsers = await client.get("http://localhost:11235/monitor/browsers")
# Endpoint statistics
stats = await client.get("http://localhost:11235/monitor/endpoints/stats")
```
- **⚡ WebSocket Streaming**: Real-time updates every 2 seconds for custom dashboards
- **🔥 Smart Browser Pool**: 3-tier architecture (permanent/hot/cold) with automatic promotion and cleanup
- **🧹 Janitor System**: Automatic resource management with event logging
- **🎮 Control Actions**: Manual browser management (kill, restart, cleanup) via API
- **📈 Production Metrics**: 6 critical metrics for operational excellence with Prometheus integration
- **🐛 Critical Bug Fixes**:
- Fixed async LLM extraction blocking issue (#1055)
- Enhanced DFS deep crawl strategy (#1607)
- Fixed sitemap parsing in AsyncUrlSeeder (#1598)
- Resolved browser viewport configuration (#1495)
- Fixed CDP timing with exponential backoff (#1528)
- Security update for pyOpenSSL (>=25.3.0)
[Full v0.7.7 Release Notes →](https://github.com/unclecode/crawl4ai/blob/main/docs/blog/release-v0.7.7.md)
</details>
<details>
<summary><strong>Version 0.7.5 Release Highlights - The Docker Hooks & Security Update</strong></summary>
@@ -977,11 +1034,14 @@ Our enterprise sponsors and technology partners help scale Crawl4AI to power pro
| Company | About | Sponsorship Tier |
|------|------|----------------------------|
| <a href="https://dashboard.capsolver.com/passport/register?inviteCode=ESVSECTX5Q23" target="_blank"><picture><source width="120" media="(prefers-color-scheme: dark)" srcset="https://docs.crawl4ai.com/uploads/sponsors/20251013045338_72a71fa4ee4d2f40.png"><source width="120" media="(prefers-color-scheme: light)" srcset="https://www.capsolver.com/assets/images/logo-text.png"><img alt="Capsolver" src="https://www.capsolver.com/assets/images/logo-text.png"></picture></a> | AI-powered Captcha solving service. Supports all major Captcha types, including reCAPTCHA, Cloudflare, and more | 🥈 Silver |
| <a href="https://app.scrapeless.com/passport/register?utm_source=official&utm_term=crawl4ai" target="_blank"><picture><source width="250" media="(prefers-color-scheme: dark)" srcset="https://gist.githubusercontent.com/aravindkarnam/0d275b942705604263e5c32d2db27bc1/raw/Scrapeless-light-logo.svg"><source width="250" media="(prefers-color-scheme: light)" srcset="https://gist.githubusercontent.com/aravindkarnam/22d0525cc0f3021bf19ebf6e11a69ccd/raw/Scrapeless-dark-logo.svg"><img alt="Scrapeless" src="https://gist.githubusercontent.com/aravindkarnam/22d0525cc0f3021bf19ebf6e11a69ccd/raw/Scrapeless-dark-logo.svg"></picture></a> | Scrapeless is the best full-stack web scraping toolkit offering Scraping API, Scraping Browser, Web Unlocker, Captcha Solver, and Proxies, designed to handle all your data collection needs. | 🥈 Silver |
| <a href="https://dashboard.capsolver.com/passport/register?inviteCode=ESVSECTX5Q23" target="_blank"><picture><source width="120" media="(prefers-color-scheme: dark)" srcset="https://docs.crawl4ai.com/uploads/sponsors/20251013045338_72a71fa4ee4d2f40.png"><source width="120" media="(prefers-color-scheme: light)" srcset="https://www.capsolver.com/assets/images/logo-text.png"><img alt="Capsolver" src="https://www.capsolver.com/assets/images/logo-text.png"></picture></a> | AI-powered Captcha solving service. Supports all major Captcha types, including reCAPTCHA, Cloudflare, and more | 🥉 Bronze |
| <a href="https://kipo.ai" target="_blank"><img src="https://docs.crawl4ai.com/uploads/sponsors/20251013045751_2d54f57f117c651e.png" alt="DataSync" width="120"/></a> | Helps engineers and buyers find, compare, and source electronic & industrial parts in seconds, with specs, pricing, lead times & alternatives.| 🥇 Gold |
| <a href="https://www.kidocode.com/" target="_blank"><img src="https://docs.crawl4ai.com/uploads/sponsors/20251013045045_bb8dace3f0440d65.svg" alt="Kidocode" width="120"/><p align="center">KidoCode</p></a> | Kidocode is a hybrid technology and entrepreneurship school for kids aged 518, offering both online and on-campus education. | 🥇 Gold |
| <a href="https://www.alephnull.sg/" target="_blank"><img src="https://docs.crawl4ai.com/uploads/sponsors/20251013050323_a9e8e8c4c3650421.svg" alt="Aleph null" width="120"/></a> | Singapore-based Aleph Null is Asias leading edtech hub, dedicated to student-centric, AI-driven education—empowering learners with the tools to thrive in a fast-changing world. | 🥇 Gold |
### 🧑‍🤝 Individual Sponsors
A heartfelt thanks to our individual supporters! Every contribution helps us keep our opensource mission alive and thriving!

View File

@@ -1,7 +1,7 @@
# crawl4ai/__version__.py
# This is the version that will be used for stable releases
__version__ = "0.7.6"
__version__ = "0.7.7"
# For nightly builds, this gets set during build process
__nightly_version__ = None

View File

@@ -1,6 +1,7 @@
import os
from typing import Union
import warnings
import requests
from .config import (
DEFAULT_PROVIDER,
DEFAULT_PROVIDER_API_KEY,
@@ -649,6 +650,85 @@ class BrowserConfig:
return config
return BrowserConfig.from_kwargs(config)
def set_nstproxy(
self,
token: str,
channel_id: str,
country: str = "ANY",
state: str = "",
city: str = "",
protocol: str = "http",
session_duration: int = 10,
):
"""
Fetch a proxy from NSTProxy API and automatically assign it to proxy_config.
Get your NSTProxy token from: https://app.nstproxy.com/profile
Args:
token (str): NSTProxy API token.
channel_id (str): NSTProxy channel ID.
country (str, optional): Country code (default: "ANY").
state (str, optional): State code (default: "").
city (str, optional): City name (default: "").
protocol (str, optional): Proxy protocol ("http" or "socks5"). Defaults to "http".
session_duration (int, optional): Session duration in minutes (0 = rotate each request). Defaults to 10.
Raises:
ValueError: If the API response format is invalid.
PermissionError: If the API returns an error message.
"""
# --- Validate input early ---
if not token or not channel_id:
raise ValueError("[NSTProxy] token and channel_id are required")
if protocol not in ("http", "socks5"):
raise ValueError(f"[NSTProxy] Invalid protocol: {protocol}")
# --- Build NSTProxy API URL ---
params = {
"fType": 2,
"count": 1,
"channelId": channel_id,
"country": country,
"protocol": protocol,
"sessionDuration": session_duration,
"token": token,
}
if state:
params["state"] = state
if city:
params["city"] = city
url = "https://api.nstproxy.com/api/v1/generate/apiproxies"
try:
response = requests.get(url, params=params, timeout=10)
response.raise_for_status()
data = response.json()
# --- Handle API error response ---
if isinstance(data, dict) and data.get("err"):
raise PermissionError(f"[NSTProxy] API Error: {data.get('msg', 'Unknown error')}")
if not isinstance(data, list) or not data:
raise ValueError("[NSTProxy] Invalid API response — expected a non-empty list")
proxy_info = data[0]
# --- Apply proxy config ---
self.proxy_config = ProxyConfig(
server=f"{protocol}://{proxy_info['ip']}:{proxy_info['port']}",
username=proxy_info["username"],
password=proxy_info["password"],
)
except Exception as e:
print(f"[NSTProxy] ❌ Failed to set proxy: {e}")
raise
class VirtualScrollConfig:
"""Configuration for virtual scroll handling.

View File

@@ -1383,9 +1383,10 @@ class AsyncPlaywrightCrawlerStrategy(AsyncCrawlerStrategy):
try:
await self.adapter.evaluate(page,
f"""
(() => {{
(async () => {{
try {{
{remove_overlays_js}
const removeOverlays = {remove_overlays_js};
await removeOverlays();
return {{ success: true }};
}} catch (error) {{
return {{

View File

@@ -845,6 +845,15 @@ class AsyncUrlSeeder:
return
data = gzip.decompress(r.content) if url.endswith(".gz") else r.content
base_url = str(r.url)
def _normalize_loc(raw: Optional[str]) -> Optional[str]:
if not raw:
return None
normalized = urljoin(base_url, raw.strip())
if not normalized:
return None
return normalized
# Detect if this is a sitemap index by checking for <sitemapindex> or presence of <sitemap> elements
is_sitemap_index = False
@@ -857,25 +866,42 @@ class AsyncUrlSeeder:
# Use XML parser for sitemaps, not HTML parser
parser = etree.XMLParser(recover=True)
root = etree.fromstring(data, parser=parser)
# Namespace-agnostic lookups using local-name() so we honor custom or missing namespaces
sitemap_loc_nodes = root.xpath("//*[local-name()='sitemap']/*[local-name()='loc']")
url_loc_nodes = root.xpath("//*[local-name()='url']/*[local-name()='loc']")
# Define namespace for sitemap
ns = {'s': 'http://www.sitemaps.org/schemas/sitemap/0.9'}
self._log(
"debug",
"Parsed sitemap {url}: {sitemap_count} sitemap entries, {url_count} url entries discovered",
params={
"url": url,
"sitemap_count": len(sitemap_loc_nodes),
"url_count": len(url_loc_nodes),
},
tag="URL_SEED",
)
# Check for sitemap index entries
sitemap_locs = root.xpath('//s:sitemap/s:loc', namespaces=ns)
if sitemap_locs:
if sitemap_loc_nodes:
is_sitemap_index = True
for sitemap_elem in sitemap_locs:
loc = sitemap_elem.text.strip() if sitemap_elem.text else ""
for sitemap_elem in sitemap_loc_nodes:
loc = _normalize_loc(sitemap_elem.text)
if loc:
sub_sitemaps.append(loc)
# If not a sitemap index, get regular URLs
if not is_sitemap_index:
for loc_elem in root.xpath('//s:url/s:loc', namespaces=ns):
loc = loc_elem.text.strip() if loc_elem.text else ""
for loc_elem in url_loc_nodes:
loc = _normalize_loc(loc_elem.text)
if loc:
regular_urls.append(loc)
if not regular_urls:
self._log(
"warning",
"No <loc> entries found inside <url> tags for sitemap {url}. The sitemap might be empty or use an unexpected structure.",
params={"url": url},
tag="URL_SEED",
)
except Exception as e:
self._log("error", "LXML parsing error for sitemap {url}: {error}",
params={"url": url, "error": str(e)}, tag="URL_SEED")
@@ -892,19 +918,39 @@ class AsyncUrlSeeder:
# Check for sitemap index entries
sitemaps = root.findall('.//sitemap')
url_entries = root.findall('.//url')
self._log(
"debug",
"ElementTree parsed sitemap {url}: {sitemap_count} sitemap entries, {url_count} url entries discovered",
params={
"url": url,
"sitemap_count": len(sitemaps),
"url_count": len(url_entries),
},
tag="URL_SEED",
)
if sitemaps:
is_sitemap_index = True
for sitemap in sitemaps:
loc_elem = sitemap.find('loc')
if loc_elem is not None and loc_elem.text:
sub_sitemaps.append(loc_elem.text.strip())
loc = _normalize_loc(loc_elem.text if loc_elem is not None else None)
if loc:
sub_sitemaps.append(loc)
# If not a sitemap index, get regular URLs
if not is_sitemap_index:
for url_elem in root.findall('.//url'):
for url_elem in url_entries:
loc_elem = url_elem.find('loc')
if loc_elem is not None and loc_elem.text:
regular_urls.append(loc_elem.text.strip())
loc = _normalize_loc(loc_elem.text if loc_elem is not None else None)
if loc:
regular_urls.append(loc)
if not regular_urls:
self._log(
"warning",
"No <loc> entries found inside <url> tags for sitemap {url}. The sitemap might be empty or use an unexpected structure.",
params={"url": url},
tag="URL_SEED",
)
except Exception as e:
self._log("error", "ElementTree parsing error for sitemap {url}: {error}",
params={"url": url, "error": str(e)}, tag="URL_SEED")

View File

@@ -617,7 +617,17 @@ class AsyncWebCrawler:
else config.chunking_strategy
)
sections = chunking.chunk(content)
extracted_content = config.extraction_strategy.run(url, sections)
# extracted_content = config.extraction_strategy.run(url, sections)
# Use async version if available for better parallelism
if hasattr(config.extraction_strategy, 'arun'):
extracted_content = await config.extraction_strategy.arun(url, sections)
else:
# Fallback to sync version run in thread pool to avoid blocking
extracted_content = await asyncio.to_thread(
config.extraction_strategy.run, url, sections
)
extracted_content = json.dumps(
extracted_content, indent=4, default=str, ensure_ascii=False
)

View File

@@ -369,6 +369,9 @@ class ManagedBrowser:
]
if self.headless:
flags.append("--headless=new")
# Add viewport flag if specified in config
if self.browser_config.viewport_height and self.browser_config.viewport_width:
flags.append(f"--window-size={self.browser_config.viewport_width},{self.browser_config.viewport_height}")
# merge common launch flags
flags.extend(self.build_browser_flags(self.browser_config))
elif self.browser_type == "firefox":
@@ -658,6 +661,11 @@ class BrowserManager:
if self.config.cdp_url or self.config.use_managed_browser:
self.config.use_managed_browser = True
cdp_url = await self.managed_browser.start() if not self.config.cdp_url else self.config.cdp_url
# Add CDP endpoint verification before connecting
if not await self._verify_cdp_ready(cdp_url):
raise Exception(f"CDP endpoint at {cdp_url} is not ready after startup")
self.browser = await self.playwright.chromium.connect_over_cdp(cdp_url)
contexts = self.browser.contexts
if contexts:
@@ -678,6 +686,24 @@ class BrowserManager:
self.default_context = self.browser
async def _verify_cdp_ready(self, cdp_url: str) -> bool:
"""Verify CDP endpoint is ready with exponential backoff"""
import aiohttp
self.logger.debug(f"Starting CDP verification for {cdp_url}", tag="BROWSER")
for attempt in range(5):
try:
async with aiohttp.ClientSession() as session:
async with session.get(f"{cdp_url}/json/version", timeout=aiohttp.ClientTimeout(total=2)) as response:
if response.status == 200:
self.logger.debug(f"CDP endpoint ready after {attempt + 1} attempts", tag="BROWSER")
return True
except Exception as e:
self.logger.debug(f"CDP check attempt {attempt + 1} failed: {e}", tag="BROWSER")
delay = 0.5 * (1.4 ** attempt)
self.logger.debug(f"Waiting {delay:.2f}s before next CDP check...", tag="BROWSER")
await asyncio.sleep(delay)
self.logger.debug(f"CDP verification failed after 5 attempts", tag="BROWSER")
return False
def _build_browser_args(self) -> dict:
"""Build browser launch arguments from config."""

View File

@@ -2,8 +2,6 @@ import click
import os
import sys
import time
import subprocess
import shutil
import humanize
from typing import Dict, Any, Optional, List
@@ -627,76 +625,6 @@ def cli():
pass
# Register server command group (Docker orchestration)
# Redirect to standalone 'cnode' CLI
@cli.command("server", context_settings=dict(
ignore_unknown_options=True,
allow_extra_args=True,
allow_interspersed_args=False
))
@click.pass_context
def server_cmd(ctx):
"""Manage Crawl4AI Docker server instances (deprecated - use 'cnode')
This command has been moved to a standalone CLI called 'cnode'.
For new installations, use:
curl -sSL https://crawl4ai.com/deploy.sh | bash
This redirect allows existing scripts to continue working.
Available commands: start, stop, status, scale, logs
Use 'crwl server <command> --help' for command-specific help.
"""
# Check if cnode is installed
cnode_path = shutil.which("cnode")
# Get all the args (subcommand + options)
args = ctx.args
if not cnode_path:
console.print(Panel(
"[yellow]The 'crwl server' command has been moved to a standalone CLI.[/yellow]\n\n"
"Please install 'cnode' (Crawl4AI Node Manager):\n"
"[cyan]curl -sSL https://crawl4ai.com/deploy.sh | bash[/cyan]\n\n"
"After installation, use:\n"
"[green]cnode <command>[/green] instead of [dim]crwl server <command>[/dim]\n\n"
"For backward compatibility, we're using the local version for now.",
title="Server Command Moved",
border_style="yellow"
))
# Try to use local version
try:
import sys
# Add deploy/docker to path
deploy_path = str(Path(__file__).parent.parent / 'deploy' / 'docker')
if deploy_path not in sys.path:
sys.path.insert(0, deploy_path)
from cnode_cli import cli as cnode_cli
# Forward to cnode with the args
sys.argv = ['cnode'] + args
cnode_cli(standalone_mode=False)
sys.exit(0)
except SystemExit as e:
# Normal exit from click
sys.exit(e.code if hasattr(e, 'code') else 0)
except Exception as e:
console.print(f"[red]Error: Could not find cnode or local server CLI: {e}[/red]")
console.print(f"[dim]Details: {e}[/dim]")
import traceback
console.print(f"[dim]{traceback.format_exc()}[/dim]")
sys.exit(1)
# cnode is installed - forward everything to it
try:
result = subprocess.run([cnode_path] + args, check=False)
sys.exit(result.returncode)
except Exception as e:
console.print(f"[red]Error running cnode: {e}[/red]")
sys.exit(1)
@cli.group("browser")
def browser_cmd():
"""Manage browser instances for Crawl4AI
@@ -1534,15 +1462,9 @@ def default(url: str, example: bool, browser_config: str, crawler_config: str, f
def main():
import sys
# Don't auto-insert 'crawl' if the command is recognized
if len(sys.argv) >= 2 and sys.argv[1] in cli.commands:
cli()
elif len(sys.argv) < 2:
cli()
else:
# Unknown command - insert 'crawl' for backward compat
if len(sys.argv) < 2 or sys.argv[1] not in cli.commands:
sys.argv.insert(1, "crawl")
cli()
cli()
if __name__ == "__main__":
main()

View File

@@ -542,6 +542,19 @@ class LXMLWebScrapingStrategy(ContentScrapingStrategy):
if el.tag in bypass_tags:
continue
# Skip elements inside <pre> or <code> tags where whitespace is significant
# This preserves whitespace-only spans (e.g., <span class="w"> </span>) in code blocks
is_in_code_block = False
ancestor = el.getparent()
while ancestor is not None:
if ancestor.tag in ("pre", "code"):
is_in_code_block = True
break
ancestor = ancestor.getparent()
if is_in_code_block:
continue
text_content = (el.text_content() or "").strip()
if (
len(text_content.split()) < word_count_threshold

View File

@@ -4,14 +4,26 @@ from typing import AsyncGenerator, Optional, Set, Dict, List, Tuple
from ..models import CrawlResult
from .bfs_strategy import BFSDeepCrawlStrategy # noqa
from ..types import AsyncWebCrawler, CrawlerRunConfig
from ..utils import normalize_url_for_deep_crawl
class DFSDeepCrawlStrategy(BFSDeepCrawlStrategy):
"""
Depth-First Search (DFS) deep crawling strategy.
Depth-first deep crawling with familiar BFS rules.
Inherits URL validation and link discovery from BFSDeepCrawlStrategy.
Overrides _arun_batch and _arun_stream to use a stack (LIFO) for DFS traversal.
We reuse the same filters, scoring, and page limits from :class:`BFSDeepCrawlStrategy`,
but walk the graph with a stack so we fully explore one branch before hopping to the
next. DFS also keeps its own ``_dfs_seen`` set so we can drop duplicate links at
discovery time without accidentally marking them as “already crawled”.
"""
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self._dfs_seen: Set[str] = set()
def _reset_seen(self, start_url: str) -> None:
"""Start each crawl with a clean dedupe set seeded with the root URL."""
self._dfs_seen = {start_url}
async def _arun_batch(
self,
start_url: str,
@@ -19,14 +31,19 @@ class DFSDeepCrawlStrategy(BFSDeepCrawlStrategy):
config: CrawlerRunConfig,
) -> List[CrawlResult]:
"""
Batch (non-streaming) DFS mode.
Uses a stack to traverse URLs in DFS order, aggregating CrawlResults into a list.
Crawl level-by-level but emit results at the end.
We keep a stack of ``(url, parent, depth)`` tuples, pop one at a time, and
hand it to ``crawler.arun_many`` with deep crawling disabled so we remain
in control of traversal. Every successful page bumps ``_pages_crawled`` and
seeds new stack items discovered via :meth:`link_discovery`.
"""
visited: Set[str] = set()
# Stack items: (url, parent_url, depth)
stack: List[Tuple[str, Optional[str], int]] = [(start_url, None, 0)]
depths: Dict[str, int] = {start_url: 0}
results: List[CrawlResult] = []
self._reset_seen(start_url)
while stack and not self._cancel_event.is_set():
url, parent, depth = stack.pop()
@@ -71,12 +88,16 @@ class DFSDeepCrawlStrategy(BFSDeepCrawlStrategy):
config: CrawlerRunConfig,
) -> AsyncGenerator[CrawlResult, None]:
"""
Streaming DFS mode.
Uses a stack to traverse URLs in DFS order and yields CrawlResults as they become available.
Same traversal as :meth:`_arun_batch`, but yield pages immediately.
Each popped URL is crawled, its metadata annotated, then the result gets
yielded before we even look at the next stack entry. Successful crawls
still feed :meth:`link_discovery`, keeping DFS order intact.
"""
visited: Set[str] = set()
stack: List[Tuple[str, Optional[str], int]] = [(start_url, None, 0)]
depths: Dict[str, int] = {start_url: 0}
self._reset_seen(start_url)
while stack and not self._cancel_event.is_set():
url, parent, depth = stack.pop()
@@ -108,3 +129,92 @@ class DFSDeepCrawlStrategy(BFSDeepCrawlStrategy):
for new_url, new_parent in reversed(new_links):
new_depth = depths.get(new_url, depth + 1)
stack.append((new_url, new_parent, new_depth))
async def link_discovery(
self,
result: CrawlResult,
source_url: str,
current_depth: int,
_visited: Set[str],
next_level: List[Tuple[str, Optional[str]]],
depths: Dict[str, int],
) -> None:
"""
Find the next URLs we should push onto the DFS stack.
Parameters
----------
result : CrawlResult
Output of the page we just crawled; its ``links`` block is our raw material.
source_url : str
URL of the parent page; stored so callers can track ancestry.
current_depth : int
Depth of the parent; children naturally sit at ``current_depth + 1``.
_visited : Set[str]
Present to match the BFS signature, but we rely on ``_dfs_seen`` instead.
next_level : list of tuples
The stack buffer supplied by the caller; we append new ``(url, parent)`` items here.
depths : dict
Shared depth map so future metadata tagging knows how deep each URL lives.
Notes
-----
- ``_dfs_seen`` keeps us from pushing duplicates without touching the traversal guard.
- Validation, scoring, and capacity trimming mirror the BFS version so behaviour stays consistent.
"""
next_depth = current_depth + 1
if next_depth > self.max_depth:
return
remaining_capacity = self.max_pages - self._pages_crawled
if remaining_capacity <= 0:
self.logger.info(
f"Max pages limit ({self.max_pages}) reached, stopping link discovery"
)
return
links = result.links.get("internal", [])
if self.include_external:
links += result.links.get("external", [])
seen = self._dfs_seen
valid_links: List[Tuple[str, float]] = []
for link in links:
raw_url = link.get("href")
if not raw_url:
continue
normalized_url = normalize_url_for_deep_crawl(raw_url, source_url)
if not normalized_url or normalized_url in seen:
continue
if not await self.can_process_url(raw_url, next_depth):
self.stats.urls_skipped += 1
continue
score = self.url_scorer.score(normalized_url) if self.url_scorer else 0
if score < self.score_threshold:
self.logger.debug(
f"URL {normalized_url} skipped: score {score} below threshold {self.score_threshold}"
)
self.stats.urls_skipped += 1
continue
seen.add(normalized_url)
valid_links.append((normalized_url, score))
if len(valid_links) > remaining_capacity:
if self.url_scorer:
valid_links.sort(key=lambda x: x[1], reverse=True)
valid_links = valid_links[:remaining_capacity]
self.logger.info(
f"Limiting to {remaining_capacity} URLs due to max_pages limit"
)
for url, score in valid_links:
if score:
result.metadata = result.metadata or {}
result.metadata["score"] = score
next_level.append((url, source_url))
depths[url] = next_depth

View File

@@ -94,6 +94,20 @@ class ExtractionStrategy(ABC):
extracted_content.extend(future.result())
return extracted_content
async def arun(self, url: str, sections: List[str], *q, **kwargs) -> List[Dict[str, Any]]:
"""
Async version: Process sections of text in parallel using asyncio.
Default implementation runs the sync version in a thread pool.
Subclasses can override this for true async processing.
:param url: The URL of the webpage.
:param sections: List of sections (strings) to process.
:return: A list of processed JSON blocks.
"""
import asyncio
return await asyncio.to_thread(self.run, url, sections, *q, **kwargs)
class NoExtractionStrategy(ExtractionStrategy):
"""
@@ -780,6 +794,177 @@ class LLMExtractionStrategy(ExtractionStrategy):
return extracted_content
async def aextract(self, url: str, ix: int, html: str) -> List[Dict[str, Any]]:
"""
Async version: Extract meaningful blocks or chunks from the given HTML using an LLM.
How it works:
1. Construct a prompt with variables.
2. Make an async request to the LLM using the prompt.
3. Parse the response and extract blocks or chunks.
Args:
url: The URL of the webpage.
ix: Index of the block.
html: The HTML content of the webpage.
Returns:
A list of extracted blocks or chunks.
"""
from .utils import aperform_completion_with_backoff
if self.verbose:
print(f"[LOG] Call LLM for {url} - block index: {ix}")
variable_values = {
"URL": url,
"HTML": escape_json_string(sanitize_html(html)),
}
prompt_with_variables = PROMPT_EXTRACT_BLOCKS
if self.instruction:
variable_values["REQUEST"] = self.instruction
prompt_with_variables = PROMPT_EXTRACT_BLOCKS_WITH_INSTRUCTION
if self.extract_type == "schema" and self.schema:
variable_values["SCHEMA"] = json.dumps(self.schema, indent=2)
prompt_with_variables = PROMPT_EXTRACT_SCHEMA_WITH_INSTRUCTION
if self.extract_type == "schema" and not self.schema:
prompt_with_variables = PROMPT_EXTRACT_INFERRED_SCHEMA
for variable in variable_values:
prompt_with_variables = prompt_with_variables.replace(
"{" + variable + "}", variable_values[variable]
)
try:
response = await aperform_completion_with_backoff(
self.llm_config.provider,
prompt_with_variables,
self.llm_config.api_token,
base_url=self.llm_config.base_url,
json_response=self.force_json_response,
extra_args=self.extra_args,
)
# Track usage
usage = TokenUsage(
completion_tokens=response.usage.completion_tokens,
prompt_tokens=response.usage.prompt_tokens,
total_tokens=response.usage.total_tokens,
completion_tokens_details=response.usage.completion_tokens_details.__dict__
if response.usage.completion_tokens_details
else {},
prompt_tokens_details=response.usage.prompt_tokens_details.__dict__
if response.usage.prompt_tokens_details
else {},
)
self.usages.append(usage)
# Update totals
self.total_usage.completion_tokens += usage.completion_tokens
self.total_usage.prompt_tokens += usage.prompt_tokens
self.total_usage.total_tokens += usage.total_tokens
try:
content = response.choices[0].message.content
blocks = None
if self.force_json_response:
blocks = json.loads(content)
if isinstance(blocks, dict):
if len(blocks) == 1 and isinstance(list(blocks.values())[0], list):
blocks = list(blocks.values())[0]
else:
blocks = [blocks]
elif isinstance(blocks, list):
blocks = blocks
else:
blocks = extract_xml_data(["blocks"], content)["blocks"]
blocks = json.loads(blocks)
for block in blocks:
block["error"] = False
except Exception:
parsed, unparsed = split_and_parse_json_objects(
response.choices[0].message.content
)
blocks = parsed
if unparsed:
blocks.append(
{"index": 0, "error": True, "tags": ["error"], "content": unparsed}
)
if self.verbose:
print(
"[LOG] Extracted",
len(blocks),
"blocks from URL:",
url,
"block index:",
ix,
)
return blocks
except Exception as e:
if self.verbose:
print(f"[LOG] Error in LLM extraction: {e}")
return [
{
"index": ix,
"error": True,
"tags": ["error"],
"content": str(e),
}
]
async def arun(self, url: str, sections: List[str]) -> List[Dict[str, Any]]:
"""
Async version: Process sections with true parallelism using asyncio.gather.
Args:
url: The URL of the webpage.
sections: List of sections (strings) to process.
Returns:
A list of extracted blocks or chunks.
"""
import asyncio
merged_sections = self._merge(
sections,
self.chunk_token_threshold,
overlap=int(self.chunk_token_threshold * self.overlap_rate),
)
extracted_content = []
# Create tasks for all sections to run in parallel
tasks = [
self.aextract(url, ix, sanitize_input_encode(section))
for ix, section in enumerate(merged_sections)
]
# Execute all tasks concurrently
results = await asyncio.gather(*tasks, return_exceptions=True)
# Process results
for result in results:
if isinstance(result, Exception):
if self.verbose:
print(f"Error in async extraction: {result}")
extracted_content.append(
{
"index": 0,
"error": True,
"tags": ["error"],
"content": str(result),
}
)
else:
extracted_content.extend(result)
return extracted_content
def show_usage(self) -> None:
"""Print a detailed token usage report showing total and per-request usage."""
print("\n=== Token Usage Summary ===")

View File

@@ -1,4 +1,4 @@
from pydantic import BaseModel, HttpUrl, PrivateAttr, Field
from pydantic import BaseModel, HttpUrl, PrivateAttr, Field, ConfigDict
from typing import List, Dict, Optional, Callable, Awaitable, Union, Any
from typing import AsyncGenerator
from typing import Generic, TypeVar
@@ -153,8 +153,7 @@ class CrawlResult(BaseModel):
console_messages: Optional[List[Dict[str, Any]]] = None
tables: List[Dict] = Field(default_factory=list) # NEW [{headers,rows,caption,summary}]
class Config:
arbitrary_types_allowed = True
model_config = ConfigDict(arbitrary_types_allowed=True)
# NOTE: The StringCompatibleMarkdown class, custom __init__ method, property getters/setters,
# and model_dump override all exist to support a smooth transition from markdown as a string
@@ -332,8 +331,7 @@ class AsyncCrawlResponse(BaseModel):
network_requests: Optional[List[Dict[str, Any]]] = None
console_messages: Optional[List[Dict[str, Any]]] = None
class Config:
arbitrary_types_allowed = True
model_config = ConfigDict(arbitrary_types_allowed=True)
###############################
# Scraping Models

View File

@@ -1,479 +0,0 @@
"""
Crawl4AI Server CLI Commands
Provides `crwl server` command group for Docker orchestration.
"""
import click
import anyio
from rich.console import Console
from rich.table import Table
from rich.panel import Panel
from rich.prompt import Confirm
from crawl4ai.server_manager import ServerManager
console = Console()
@click.group("server")
def server_cmd():
"""Manage Crawl4AI Docker server instances
One-command deployment with automatic scaling:
- Single container for development (N=1)
- Docker Swarm for production with built-in load balancing (N>1)
- Docker Compose + Nginx as fallback (N>1)
Examples:
crwl server start # Single container on port 11235
crwl server start --replicas 3 # Auto-detect Swarm or Compose
crwl server start -r 5 --port 8080 # 5 replicas on custom port
crwl server status # Check current deployment
crwl server scale 10 # Scale to 10 replicas
crwl server stop # Stop and cleanup
"""
pass
@server_cmd.command("start")
@click.option(
"--replicas", "-r",
type=int,
default=1,
help="Number of container replicas (default: 1)"
)
@click.option(
"--mode",
type=click.Choice(["auto", "single", "swarm", "compose"]),
default="auto",
help="Deployment mode (default: auto-detect)"
)
@click.option(
"--port", "-p",
type=int,
default=11235,
help="External port to expose (default: 11235)"
)
@click.option(
"--env-file",
type=click.Path(exists=True),
help="Path to environment file"
)
@click.option(
"--image",
default="unclecode/crawl4ai:latest",
help="Docker image to use (default: unclecode/crawl4ai:latest)"
)
def start_cmd(replicas: int, mode: str, port: int, env_file: str, image: str):
"""Start Crawl4AI server with automatic orchestration.
Deployment modes:
- auto: Automatically choose best mode (default)
- single: Single container (N=1 only)
- swarm: Docker Swarm with built-in load balancing
- compose: Docker Compose + Nginx reverse proxy
The server will:
1. Check if Docker is running
2. Validate port availability
3. Pull image if needed
4. Start container(s) with health checks
5. Save state for management
Examples:
# Development: single container
crwl server start
# Production: 5 replicas with Swarm
crwl server start --replicas 5
# Custom configuration
crwl server start -r 3 --port 8080 --env-file .env.prod
"""
manager = ServerManager()
console.print(Panel(
f"[cyan]Starting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{replicas}[/yellow]\n"
f"Mode: [yellow]{mode}[/yellow]\n"
f"Port: [yellow]{port}[/yellow]\n"
f"Image: [yellow]{image}[/yellow]",
title="Server Start",
border_style="cyan"
))
with console.status("[cyan]Starting server..."):
async def _start():
return await manager.start(
replicas=replicas,
mode=mode,
port=port,
env_file=env_file,
image=image
)
result = anyio.run(_start)
if result["success"]:
console.print(Panel(
f"[green]✓ Server started successfully![/green]\n\n"
f"Mode: [cyan]{result.get('state_data', {}).get('mode', mode)}[/cyan]\n"
f"URL: [bold]http://localhost:{port}[/bold]\n"
f"Health: [bold]http://localhost:{port}/health[/bold]\n"
f"Monitor: [bold]http://localhost:{port}/monitor[/bold]",
title="Server Running",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Failed to start server[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "already running" in error_msg.lower():
console.print("\n[yellow]Hint: Use 'crwl server status' to check current deployment[/yellow]")
console.print("[yellow] Use 'crwl server stop' to stop existing server[/yellow]")
@server_cmd.command("status")
def status_cmd():
"""Show current server status and deployment info.
Displays:
- Running state (up/down)
- Deployment mode (single/swarm/compose)
- Number of replicas
- Port mapping
- Uptime
- Image version
Example:
crwl server status
"""
manager = ServerManager()
async def _status():
return await manager.status()
result = anyio.run(_status)
if result["running"]:
table = Table(title="Crawl4AI Server Status", border_style="green")
table.add_column("Property", style="cyan")
table.add_column("Value", style="green")
table.add_row("Status", "🟢 Running")
table.add_row("Mode", result["mode"])
table.add_row("Replicas", str(result.get("replicas", 1)))
table.add_row("Port", str(result.get("port", 11235)))
table.add_row("Image", result.get("image", "unknown"))
table.add_row("Uptime", result.get("uptime", "unknown"))
table.add_row("Started", result.get("started_at", "unknown"))
console.print(table)
console.print(f"\n[green]✓ Server is healthy[/green]")
console.print(f"[dim]Access: http://localhost:{result.get('port', 11235)}[/dim]")
else:
console.print(Panel(
f"[yellow]No server is currently running[/yellow]\n\n"
f"Use 'crwl server start' to launch a server",
title="Server Status",
border_style="yellow"
))
@server_cmd.command("stop")
@click.option(
"--remove-volumes",
is_flag=True,
help="Remove associated volumes (WARNING: deletes data)"
)
def stop_cmd(remove_volumes: bool):
"""Stop running Crawl4AI server and cleanup resources.
This will:
1. Stop all running containers/services
2. Remove containers
3. Optionally remove volumes (--remove-volumes)
4. Clean up state files
WARNING: Use --remove-volumes with caution as it will delete
persistent data including Redis databases and logs.
Examples:
# Stop server, keep volumes
crwl server stop
# Stop and remove all data
crwl server stop --remove-volumes
"""
manager = ServerManager()
# Confirm if removing volumes
if remove_volumes:
if not Confirm.ask(
"[red]⚠️ This will delete all server data including Redis databases. Continue?[/red]"
):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Stopping server..."):
async def _stop():
return await manager.stop(remove_volumes=remove_volumes)
result = anyio.run(_stop)
if result["success"]:
console.print(Panel(
f"[green]✓ Server stopped successfully[/green]\n\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Server Stopped",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Error stopping server[/red]\n\n"
f"{result.get('error', result.get('message', 'Unknown error'))}",
title="Error",
border_style="red"
))
@server_cmd.command("scale")
@click.argument("replicas", type=int)
def scale_cmd(replicas: int):
"""Scale server to specified number of replicas.
Only works with Swarm or Compose modes. Single container
mode cannot be scaled (must stop and restart with --replicas).
Scaling is live and does not require downtime. The load
balancer will automatically distribute traffic to new replicas.
Examples:
# Scale up to 10 replicas
crwl server scale 10
# Scale down to 2 replicas
crwl server scale 2
# Scale to 1 (minimum)
crwl server scale 1
"""
if replicas < 1:
console.print("[red]Error: Replicas must be at least 1[/red]")
return
manager = ServerManager()
with console.status(f"[cyan]Scaling to {replicas} replicas..."):
async def _scale():
return await manager.scale(replicas=replicas)
result = anyio.run(_scale)
if result["success"]:
console.print(Panel(
f"[green]✓ Scaled successfully[/green]\n\n"
f"New replica count: [bold]{replicas}[/bold]\n"
f"Mode: [cyan]{result.get('mode')}[/cyan]",
title="Scaling Complete",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Scaling failed[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "single container" in error_msg.lower():
console.print("\n[yellow]Hint: For single container mode:[/yellow]")
console.print("[yellow] 1. crwl server stop[/yellow]")
console.print(f"[yellow] 2. crwl server start --replicas {replicas}[/yellow]")
@server_cmd.command("logs")
@click.option(
"--follow", "-f",
is_flag=True,
help="Follow log output (like tail -f)"
)
@click.option(
"--tail",
type=int,
default=100,
help="Number of lines to show (default: 100)"
)
def logs_cmd(follow: bool, tail: int):
"""View server logs.
Shows logs from running containers/services. Use --follow
to stream logs in real-time.
Examples:
# Show last 100 lines
crwl server logs
# Show last 500 lines
crwl server logs --tail 500
# Follow logs in real-time
crwl server logs --follow
# Combine options
crwl server logs -f --tail 50
"""
manager = ServerManager()
async def _logs():
return await manager.logs(follow=follow, tail=tail)
output = anyio.run(_logs)
console.print(output)
@server_cmd.command("cleanup")
@click.option(
"--force",
is_flag=True,
help="Force cleanup even if state file doesn't exist"
)
def cleanup_cmd(force: bool):
"""Force cleanup of all Crawl4AI Docker resources.
Stops and removes all containers, networks, and optionally volumes.
Useful when server is stuck or state is corrupted.
Examples:
# Clean up everything
crwl server cleanup
# Force cleanup (ignore state file)
crwl server cleanup --force
"""
manager = ServerManager()
console.print(Panel(
f"[yellow]⚠️ Cleaning up Crawl4AI Docker resources[/yellow]\n\n"
f"This will stop and remove:\n"
f"- All Crawl4AI containers\n"
f"- Nginx load balancer\n"
f"- Redis instance\n"
f"- Docker networks\n"
f"- State files",
title="Cleanup",
border_style="yellow"
))
if not force and not Confirm.ask("[yellow]Continue with cleanup?[/yellow]"):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Cleaning up resources..."):
async def _cleanup():
return await manager.cleanup(force=force)
result = anyio.run(_cleanup)
if result["success"]:
console.print(Panel(
f"[green]✓ Cleanup completed successfully[/green]\n\n"
f"Removed: {result.get('removed', 0)} containers\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Cleanup Complete",
border_style="green"
))
else:
console.print(Panel(
f"[yellow]⚠️ Partial cleanup[/yellow]\n\n"
f"{result.get('message', 'Some resources may still exist')}",
title="Cleanup Status",
border_style="yellow"
))
@server_cmd.command("restart")
@click.option(
"--replicas", "-r",
type=int,
help="New replica count (optional)"
)
def restart_cmd(replicas: int):
"""Restart server (stop then start with same config).
Preserves existing configuration unless overridden with options.
Useful for applying image updates or recovering from errors.
Examples:
# Restart with same configuration
crwl server restart
# Restart and change replica count
crwl server restart --replicas 5
"""
manager = ServerManager()
# Get current state
async def _get_status():
return await manager.status()
current = anyio.run(_get_status)
if not current["running"]:
console.print("[yellow]No server is running. Use 'crwl server start' instead.[/yellow]")
return
# Extract current config
current_replicas = current.get("replicas", 1)
current_port = current.get("port", 11235)
current_image = current.get("image", "unclecode/crawl4ai:latest")
current_mode = current.get("mode", "auto")
# Override with CLI args
new_replicas = replicas if replicas is not None else current_replicas
console.print(Panel(
f"[cyan]Restarting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{current_replicas}[/yellow] → [green]{new_replicas}[/green]\n"
f"Port: [yellow]{current_port}[/yellow]\n"
f"Mode: [yellow]{current_mode}[/yellow]",
title="Server Restart",
border_style="cyan"
))
# Stop current
with console.status("[cyan]Stopping current server..."):
async def _stop_server():
return await manager.stop(remove_volumes=False)
stop_result = anyio.run(_stop_server)
if not stop_result["success"]:
console.print(f"[red]Failed to stop server: {stop_result.get('error')}[/red]")
return
# Start new
with console.status("[cyan]Starting server..."):
async def _start_server():
return await manager.start(
replicas=new_replicas,
mode="auto",
port=current_port,
image=current_image
)
start_result = anyio.run(_start_server)
if start_result["success"]:
console.print(Panel(
f"[green]✓ Server restarted successfully![/green]\n\n"
f"URL: [bold]http://localhost:{current_port}[/bold]",
title="Restart Complete",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Failed to restart server[/red]\n\n"
f"{start_result.get('error', 'Unknown error')}",
title="Error",
border_style="red"
))

File diff suppressed because it is too large Load Diff

View File

@@ -1,52 +0,0 @@
version: '3.8'
services:
redis:
image: redis:alpine
command: redis-server --appendonly yes
volumes:
- redis_data:/data
networks:
- crawl4ai_net
restart: unless-stopped
crawl4ai:
image: ${IMAGE}
deploy:
replicas: ${REPLICAS}
resources:
limits:
memory: 4G
shm_size: 1g
environment:
- REDIS_HOST=redis
- REDIS_PORT=6379
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:11235/health"]
interval: 30s
timeout: 10s
retries: 3
start_period: 40s
depends_on:
- redis
networks:
- crawl4ai_net
nginx:
image: nginx:alpine
ports:
- "${PORT}:80"
volumes:
- ${NGINX_CONF}:/etc/nginx/nginx.conf:ro
depends_on:
- crawl4ai
networks:
- crawl4ai_net
restart: unless-stopped
networks:
crawl4ai_net:
driver: bridge
volumes:
redis_data:

View File

@@ -1,75 +0,0 @@
events {
worker_connections 1024;
}
http {
upstream crawl4ai_backend {
# DNS-based load balancing to Docker Compose service
# Docker Compose provides DNS resolution for service name
server crawl4ai:11235 max_fails=3 fail_timeout=30s;
# Keep connections alive
keepalive 32;
}
# Sticky sessions for monitoring (same IP always goes to same container)
upstream crawl4ai_monitor {
ip_hash; # Sticky sessions based on client IP
server crawl4ai:11235 max_fails=3 fail_timeout=30s;
keepalive 32;
}
server {
listen 80;
server_name _;
# Increase timeouts for long-running crawl operations
proxy_connect_timeout 300;
proxy_send_timeout 300;
proxy_read_timeout 300;
send_timeout 300;
# WebSocket endpoint for real-time monitoring (exact match)
location = /monitor/ws {
proxy_pass http://crawl4ai_monitor/monitor/ws;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
# WebSocket timeouts
proxy_connect_timeout 7d;
proxy_send_timeout 7d;
proxy_read_timeout 7d;
}
# Monitor and dashboard with sticky sessions (regex location)
location ~ ^/(monitor|dashboard) {
proxy_pass http://crawl4ai_monitor;
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Forwarded-Proto $scheme;
}
# HTTP endpoints (load balanced)
location / {
proxy_pass http://crawl4ai_backend;
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Forwarded-Proto $scheme;
# Support large request bodies (for batch operations)
client_max_body_size 10M;
}
# Health check endpoint (bypass load balancer)
location /health {
proxy_pass http://crawl4ai_backend/health;
access_log off;
}
}
}

View File

@@ -1825,6 +1825,82 @@ def perform_completion_with_backoff(
# ]
async def aperform_completion_with_backoff(
provider,
prompt_with_variables,
api_token,
json_response=False,
base_url=None,
**kwargs,
):
"""
Async version: Perform an API completion request with exponential backoff.
How it works:
1. Sends an async completion request to the API.
2. Retries on rate-limit errors with exponential delays (async).
3. Returns the API response or an error after all retries.
Args:
provider (str): The name of the API provider.
prompt_with_variables (str): The input prompt for the completion request.
api_token (str): The API token for authentication.
json_response (bool): Whether to request a JSON response. Defaults to False.
base_url (Optional[str]): The base URL for the API. Defaults to None.
**kwargs: Additional arguments for the API request.
Returns:
dict: The API response or an error message after all retries.
"""
from litellm import acompletion
from litellm.exceptions import RateLimitError
import asyncio
max_attempts = 3
base_delay = 2 # Base delay in seconds, you can adjust this based on your needs
extra_args = {"temperature": 0.01, "api_key": api_token, "base_url": base_url}
if json_response:
extra_args["response_format"] = {"type": "json_object"}
if kwargs.get("extra_args"):
extra_args.update(kwargs["extra_args"])
for attempt in range(max_attempts):
try:
response = await acompletion(
model=provider,
messages=[{"role": "user", "content": prompt_with_variables}],
**extra_args,
)
return response # Return the successful response
except RateLimitError as e:
print("Rate limit error:", str(e))
if attempt == max_attempts - 1:
# Last attempt failed, raise the error.
raise
# Check if we have exhausted our max attempts
if attempt < max_attempts - 1:
# Calculate the delay and wait
delay = base_delay * (2**attempt) # Exponential backoff formula
print(f"Waiting for {delay} seconds before retrying...")
await asyncio.sleep(delay)
else:
# Return an error response after exhausting all retries
return [
{
"index": 0,
"tags": ["error"],
"content": ["Rate limit error. Please try again later."],
}
]
except Exception as e:
raise e # Raise any other exceptions immediately
def extract_blocks(url, html, provider=DEFAULT_PROVIDER, api_token=None, base_url=None):
"""
Extract content blocks from website HTML using an AI provider.

View File

@@ -1,402 +0,0 @@
# Crawl4AI DevOps Agent Context
## Service Overview
**Crawl4AI**: Browser-based web crawling service with AI extraction. Docker deployment with horizontal scaling (1-N containers), Redis coordination, Nginx load balancing.
## Architecture Quick Reference
```
Client → Nginx:11235 → [crawl4ai-1, crawl4ai-2, ...crawl4ai-N] ← Redis
Monitor Dashboard
```
**Components:**
- **Nginx**: Load balancer (round-robin API, sticky monitoring)
- **Crawl4AI containers**: FastAPI + Playwright browsers
- **Redis**: Container discovery (heartbeats 30s), monitoring data aggregation
- **Monitor**: Real-time dashboard at `/dashboard`
## CLI Commands
### Start/Stop
```bash
crwl server start [-r N] [--port P] [--mode auto|single|swarm|compose] [--env-file F] [--image I]
crwl server stop [--remove-volumes]
crwl server restart [-r N]
```
### Management
```bash
crwl server status # Show mode, replicas, port, uptime
crwl server scale N # Live scaling (Swarm/Compose only)
crwl server logs [-f] [--tail N]
```
**Defaults**: replicas=1, port=11235, mode=auto, image=unclecode/crawl4ai:latest
## Deployment Modes
| Replicas | Mode | Load Balancer | Use Case |
|----------|------|---------------|----------|
| N=1 | single | None | Dev/testing |
| N>1 | swarm | Built-in | Production (if `docker swarm init` done) |
| N>1 | compose | Nginx | Production (fallback) |
**Mode Detection** (when mode=auto):
1. If N=1 → single
2. If N>1 & Swarm active → swarm
3. If N>1 & Swarm inactive → compose
## File Locations
```
~/.crawl4ai/server/
├── state.json # Current deployment state
├── docker-compose.yml # Generated compose file
└── nginx.conf # Generated nginx config
/app/ # Inside container
├── deploy/docker/server.py
├── deploy/docker/monitor.py
├── deploy/docker/static/monitor/index.html
└── crawler_pool.py # Browser pool (PERMANENT, HOT_POOL, COLD_POOL)
```
## Monitoring & Troubleshooting
### Health Checks
```bash
curl http://localhost:11235/health # Service health
curl http://localhost:11235/monitor/containers # Container discovery
curl http://localhost:11235/monitor/requests # Aggregated requests
```
### Dashboard
- URL: `http://localhost:11235/dashboard/`
- Features: Container filtering (All/C-1/C-2/C-3), real-time WebSocket, timeline charts
- WebSocket: `/monitor/ws` (sticky sessions)
### Common Issues
**No containers showing in dashboard:**
```bash
docker exec <redis-container> redis-cli SMEMBERS monitor:active_containers
docker exec <redis-container> redis-cli KEYS "monitor:heartbeat:*"
```
Wait 30s for heartbeat registration.
**Load balancing not working:**
```bash
docker exec <nginx-container> cat /etc/nginx/nginx.conf | grep upstream
docker logs <nginx-container> | grep error
```
Check Nginx upstream has no `ip_hash` for API endpoints.
**Redis connection errors:**
```bash
docker logs <crawl4ai-container> | grep -i redis
docker exec <crawl4ai-container> ping redis
```
Verify REDIS_HOST=redis, REDIS_PORT=6379.
**Containers not scaling:**
```bash
# Swarm
docker service ls
docker service ps crawl4ai
# Compose
docker compose -f ~/.crawl4ai/server/docker-compose.yml ps
docker compose -f ~/.crawl4ai/server/docker-compose.yml up -d --scale crawl4ai=N
```
### Redis Data Structure
```
monitor:active_containers # SET: {container_ids}
monitor:heartbeat:{cid} # STRING: {id, hostname, last_seen} TTL=60s
monitor:{cid}:active_requests # STRING: JSON list, TTL=5min
monitor:{cid}:completed # STRING: JSON list, TTL=1h
monitor:{cid}:janitor # STRING: JSON list, TTL=1h
monitor:{cid}:errors # STRING: JSON list, TTL=1h
monitor:endpoint_stats # STRING: JSON aggregate, TTL=24h
```
## Environment Variables
### Required for Multi-LLM
```bash
OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
DEEPSEEK_API_KEY=...
GROQ_API_KEY=...
TOGETHER_API_KEY=...
MISTRAL_API_KEY=...
GEMINI_API_TOKEN=...
```
### Redis Configuration (Optional)
```bash
REDIS_HOST=redis # Default: redis
REDIS_PORT=6379 # Default: 6379
REDIS_TTL_ACTIVE_REQUESTS=300 # Default: 5min
REDIS_TTL_COMPLETED_REQUESTS=3600 # Default: 1h
REDIS_TTL_JANITOR_EVENTS=3600 # Default: 1h
REDIS_TTL_ERRORS=3600 # Default: 1h
REDIS_TTL_ENDPOINT_STATS=86400 # Default: 24h
REDIS_TTL_HEARTBEAT=60 # Default: 1min
```
## API Endpoints
### Core API
- `POST /crawl` - Crawl URL (load-balanced)
- `POST /batch` - Batch crawl (load-balanced)
- `GET /health` - Health check (load-balanced)
### Monitor API (Aggregated from all containers)
- `GET /monitor/health` - Local container health
- `GET /monitor/containers` - All active containers
- `GET /monitor/requests` - All requests (active + completed)
- `GET /monitor/browsers` - Browser pool status (local only)
- `GET /monitor/logs/janitor` - Janitor cleanup events
- `GET /monitor/logs/errors` - Error logs
- `GET /monitor/endpoints/stats` - Endpoint analytics
- `WS /monitor/ws` - Real-time updates (aggregated)
### Control Actions
- `POST /monitor/actions/cleanup` - Force browser cleanup
- `POST /monitor/actions/kill_browser` - Kill specific browser
- `POST /monitor/actions/restart_browser` - Restart browser
- `POST /monitor/stats/reset` - Reset endpoint counters
## Docker Commands Reference
### Inspection
```bash
# List containers
docker ps --filter "name=crawl4ai"
# Container logs
docker logs <container-id> -f --tail 100
# Redis CLI
docker exec -it <redis-container> redis-cli
KEYS monitor:*
SMEMBERS monitor:active_containers
GET monitor:<cid>:completed
TTL monitor:heartbeat:<cid>
# Nginx config
docker exec <nginx-container> cat /etc/nginx/nginx.conf
# Container stats
docker stats --no-stream --format "table {{.Name}}\t{{.CPUPerc}}\t{{.MemUsage}}"
```
### Compose Operations
```bash
# Scale
docker compose -f ~/.crawl4ai/server/docker-compose.yml up -d --scale crawl4ai=5
# Restart service
docker compose -f ~/.crawl4ai/server/docker-compose.yml restart crawl4ai
# View services
docker compose -f ~/.crawl4ai/server/docker-compose.yml ps
```
### Swarm Operations
```bash
# Initialize Swarm
docker swarm init
# Scale service
docker service scale crawl4ai=5
# Service info
docker service ls
docker service ps crawl4ai --no-trunc
# Service logs
docker service logs crawl4ai --tail 100 -f
```
## Performance & Scaling
### Resource Recommendations
| Containers | Memory/Container | Total Memory | Use Case |
|------------|-----------------|--------------|----------|
| 1 | 4GB | 4GB | Development |
| 3 | 4GB | 12GB | Small prod |
| 5 | 4GB | 20GB | Medium prod |
| 10 | 4GB | 40GB | Large prod |
**Expected Throughput**: ~10 req/min per container (depends on crawl complexity)
### Scaling Guidelines
- **Horizontal**: Add replicas (`crwl server scale N`)
- **Vertical**: Adjust `--memory 8G --cpus 4` in kwargs
- **Browser Pool**: Permanent (1) + Hot pool (adaptive) + Cold pool (cleanup by janitor)
### Redis Memory Usage
- **Per container**: ~110KB (requests + events + errors + heartbeat)
- **10 containers**: ~1.1MB
- **Recommendation**: 256MB Redis is sufficient for <100 containers
## Security Notes
### Input Validation
All CLI inputs validated:
- Image name: alphanumeric + `.-/:_@` only, max 256 chars
- Port: 1-65535
- Replicas: 1-100
- Env file: must exist and be readable
- Container IDs: alphanumeric + `-_` only (prevents Redis injection)
### Network Security
- Nginx forwards to internal `crawl4ai` service (Docker network)
- Monitor endpoints have NO authentication (add MONITOR_TOKEN env for security)
- Redis is internal-only (no external port)
### Recommended Production Setup
```bash
# Add authentication
export MONITOR_TOKEN="your-secret-token"
# Use Redis password
redis:
command: redis-server --requirepass ${REDIS_PASSWORD}
# Enable rate limiting in Nginx
limit_req_zone $binary_remote_addr zone=api:10m rate=10r/s;
```
## Common User Scenarios
### Scenario 1: Fresh Deployment
```bash
crwl server start --replicas 3 --env-file .env
# Wait for health check, then access http://localhost:11235/health
```
### Scenario 2: Scaling Under Load
```bash
crwl server scale 10
# Live scaling, no downtime
```
### Scenario 3: Debugging Slow Requests
```bash
# Check dashboard
open http://localhost:11235/dashboard/
# Check container logs
docker logs <slowest-container-id> --tail 100
# Check browser pool
curl http://localhost:11235/monitor/browsers | jq
```
### Scenario 4: Redis Connection Issues
```bash
# Check Redis connectivity
docker exec <crawl4ai-container> nc -zv redis 6379
# Check Redis logs
docker logs <redis-container>
# Restart containers (triggers reconnect with retry logic)
crwl server restart
```
### Scenario 5: Container Not Appearing in Dashboard
```bash
# Wait 30s for heartbeat
sleep 30
# Check Redis
docker exec <redis-container> redis-cli SMEMBERS monitor:active_containers
# Check container logs for heartbeat errors
docker logs <missing-container> | grep -i heartbeat
```
## Code Context for Advanced Debugging
### Key Classes
- `MonitorStats` (monitor.py): Tracks stats, Redis persistence, heartbeat worker
- `ServerManager` (server_manager.py): CLI orchestration, mode detection
- Browser pool globals: `PERMANENT`, `HOT_POOL`, `COLD_POOL`, `LOCK` (crawler_pool.py)
### Critical Timeouts
- Browser pool lock: 2s timeout (prevents deadlock)
- WebSocket connection: 5s timeout
- Health check: 30-60s timeout
- Heartbeat interval: 30s, TTL: 60s
- Redis retry: 3 attempts, backoff: 0.5s/1s/2s
- Circuit breaker: 5 failures → 5min backoff
### State Transitions
```
NOT_RUNNING → STARTING → HEALTHY → RUNNING
↓ ↓
FAILED UNHEALTHY → STOPPED
```
State file: `~/.crawl4ai/server/state.json` (atomic writes, fcntl locking)
## Quick Diagnostic Commands
```bash
# Full system check
crwl server status
docker ps
curl http://localhost:11235/health
curl http://localhost:11235/monitor/containers | jq
# Redis check
docker exec <redis-container> redis-cli PING
docker exec <redis-container> redis-cli INFO stats
# Network check
docker network ls
docker network inspect <network-name>
# Logs check
docker logs <nginx-container> --tail 50
docker logs <redis-container> --tail 50
docker compose -f ~/.crawl4ai/server/docker-compose.yml logs --tail 100
```
## Agent Decision Tree
**User reports slow crawling:**
1. Check dashboard for active requests stuck → kill browser if >5min
2. Check browser pool status → cleanup if hot/cold pool >10
3. Check container CPU/memory → scale up if >80%
4. Check Redis latency → restart Redis if >100ms
**User reports missing containers:**
1. Wait 30s for heartbeat
2. Check `docker ps` vs dashboard count
3. Check Redis SMEMBERS monitor:active_containers
4. Check container logs for Redis connection errors
5. Verify REDIS_HOST/PORT env vars
**User reports 502/503 errors:**
1. Check Nginx logs for upstream errors
2. Check container health: `curl http://localhost:11235/health`
3. Check if all containers are healthy: `docker ps`
4. Restart Nginx: `docker restart <nginx-container>`
**User wants to update image:**
1. `crwl server stop`
2. `docker pull unclecode/crawl4ai:latest`
3. `crwl server start --replicas <previous-count>`
---
**Version**: Crawl4AI v0.7.4+
**Last Updated**: 2025-01-20
**AI Agent Note**: All commands, file paths, and Redis keys verified against codebase. Use exact syntax shown. For user-facing responses, translate technical details to plain language.

File diff suppressed because it is too large Load Diff

View File

@@ -59,13 +59,13 @@ Pull and run images directly from Docker Hub without building locally.
#### 1. Pull the Image
Our latest stable release is `0.7.6`. Images are built with multi-arch manifests, so Docker automatically pulls the correct version for your system.
Our latest stable release is `0.7.7`. Images are built with multi-arch manifests, so Docker automatically pulls the correct version for your system.
```bash
# Pull the latest stable version (0.7.6)
docker pull unclecode/crawl4ai:0.7.6
# Pull the latest stable version (0.7.7)
docker pull unclecode/crawl4ai:0.7.7
# Or use the latest tag (points to 0.7.6)
# Or use the latest tag (points to 0.7.7)
docker pull unclecode/crawl4ai:latest
```
@@ -100,7 +100,7 @@ EOL
-p 11235:11235 \
--name crawl4ai \
--shm-size=1g \
unclecode/crawl4ai:0.7.6
unclecode/crawl4ai:0.7.7
```
* **With LLM support:**
@@ -111,7 +111,7 @@ EOL
--name crawl4ai \
--env-file .llm.env \
--shm-size=1g \
unclecode/crawl4ai:0.7.6
unclecode/crawl4ai:0.7.7
```
> The server will be available at `http://localhost:11235`. Visit `/playground` to access the interactive testing interface.
@@ -184,7 +184,7 @@ The `docker-compose.yml` file in the project root provides a simplified approach
```bash
# Pulls and runs the release candidate from Docker Hub
# Automatically selects the correct architecture
IMAGE=unclecode/crawl4ai:0.7.6 docker compose up -d
IMAGE=unclecode/crawl4ai:0.7.7 docker compose up -d
```
* **Build and Run Locally:**

View File

@@ -1 +0,0 @@
# Deploy docker module

View File

@@ -1,492 +0,0 @@
"""
Crawl4AI Server CLI Commands
Provides `cnode` command group for Docker orchestration.
"""
import click
import anyio
from rich.console import Console
from rich.table import Table
from rich.panel import Panel
from rich.prompt import Confirm
from deploy.docker.server_manager import ServerManager
console = Console()
@click.group()
def cli():
"""Manage Crawl4AI Docker server instances
\b
One-command deployment with automatic scaling:
• Single container for development (N=1)
• Docker Swarm for production with built-in load balancing (N>1)
• Docker Compose + Nginx as fallback (N>1)
\b
Examples:
cnode start # Single container on port 11235
cnode start --replicas 3 # Auto-detect Swarm or Compose
cnode start -r 5 --port 8080 # 5 replicas on custom port
cnode status # Check current deployment
cnode scale 10 # Scale to 10 replicas
cnode stop # Stop and cleanup
"""
pass
@cli.command("start")
@click.option(
"--replicas", "-r",
type=int,
default=1,
help="Number of container replicas (default: 1)"
)
@click.option(
"--mode",
type=click.Choice(["auto", "single", "swarm", "compose"]),
default="auto",
help="Deployment mode (default: auto-detect)"
)
@click.option(
"--port", "-p",
type=int,
default=11235,
help="External port to expose (default: 11235)"
)
@click.option(
"--env-file",
type=click.Path(exists=True),
help="Path to environment file"
)
@click.option(
"--image",
default="unclecode/crawl4ai:latest",
help="Docker image to use (default: unclecode/crawl4ai:latest)"
)
def start_cmd(replicas: int, mode: str, port: int, env_file: str, image: str):
"""Start Crawl4AI server with automatic orchestration.
Deployment modes:
- auto: Automatically choose best mode (default)
- single: Single container (N=1 only)
- swarm: Docker Swarm with built-in load balancing
- compose: Docker Compose + Nginx reverse proxy
The server will:
1. Check if Docker is running
2. Validate port availability
3. Pull image if needed
4. Start container(s) with health checks
5. Save state for management
Examples:
# Development: single container
cnode start
# Production: 5 replicas with Swarm
cnode start --replicas 5
# Custom configuration
cnode start -r 3 --port 8080 --env-file .env.prod
"""
manager = ServerManager()
console.print(Panel(
f"[cyan]Starting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{replicas}[/yellow]\n"
f"Mode: [yellow]{mode}[/yellow]\n"
f"Port: [yellow]{port}[/yellow]\n"
f"Image: [yellow]{image}[/yellow]",
title="Server Start",
border_style="cyan"
))
with console.status("[cyan]Starting server..."):
async def _start():
return await manager.start(
replicas=replicas,
mode=mode,
port=port,
env_file=env_file,
image=image
)
result = anyio.run(_start)
if result["success"]:
console.print(Panel(
f"[green]✓ Server started successfully![/green]\n\n"
f"Mode: [cyan]{result.get('state_data', {}).get('mode', mode)}[/cyan]\n"
f"URL: [bold]http://localhost:{port}[/bold]\n"
f"Health: [bold]http://localhost:{port}/health[/bold]\n"
f"Monitor: [bold]http://localhost:{port}/monitor[/bold]",
title="Server Running",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Failed to start server[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "already running" in error_msg.lower():
console.print("\n[yellow]Hint: Use 'cnode status' to check current deployment[/yellow]")
console.print("[yellow] Use 'cnode stop' to stop existing server[/yellow]")
@cli.command("status")
def status_cmd():
"""Show current server status and deployment info.
Displays:
- Running state (up/down)
- Deployment mode (single/swarm/compose)
- Number of replicas
- Port mapping
- Uptime
- Image version
Example:
cnode status
"""
manager = ServerManager()
async def _status():
return await manager.status()
result = anyio.run(_status)
if result["running"]:
table = Table(title="Crawl4AI Server Status", border_style="green")
table.add_column("Property", style="cyan")
table.add_column("Value", style="green")
table.add_row("Status", "🟢 Running")
table.add_row("Mode", result["mode"])
table.add_row("Replicas", str(result.get("replicas", 1)))
table.add_row("Port", str(result.get("port", 11235)))
table.add_row("Image", result.get("image", "unknown"))
table.add_row("Uptime", result.get("uptime", "unknown"))
table.add_row("Started", result.get("started_at", "unknown"))
console.print(table)
console.print(f"\n[green]✓ Server is healthy[/green]")
console.print(f"[dim]Access: http://localhost:{result.get('port', 11235)}[/dim]")
else:
console.print(Panel(
f"[yellow]No server is currently running[/yellow]\n\n"
f"Use 'cnode start' to launch a server",
title="Server Status",
border_style="yellow"
))
@cli.command("stop")
@click.option(
"--remove-volumes",
is_flag=True,
help="Remove associated volumes (WARNING: deletes data)"
)
def stop_cmd(remove_volumes: bool):
"""Stop running Crawl4AI server and cleanup resources.
This will:
1. Stop all running containers/services
2. Remove containers
3. Optionally remove volumes (--remove-volumes)
4. Clean up state files
WARNING: Use --remove-volumes with caution as it will delete
persistent data including Redis databases and logs.
Examples:
# Stop server, keep volumes
cnode stop
# Stop and remove all data
cnode stop --remove-volumes
"""
manager = ServerManager()
# Confirm if removing volumes
if remove_volumes:
if not Confirm.ask(
"[red]⚠️ This will delete all server data including Redis databases. Continue?[/red]"
):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Stopping server..."):
async def _stop():
return await manager.stop(remove_volumes=remove_volumes)
result = anyio.run(_stop)
if result["success"]:
console.print(Panel(
f"[green]✓ Server stopped successfully[/green]\n\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Server Stopped",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Error stopping server[/red]\n\n"
f"{result.get('error', result.get('message', 'Unknown error'))}",
title="Error",
border_style="red"
))
@cli.command("scale")
@click.argument("replicas", type=int)
def scale_cmd(replicas: int):
"""Scale server to specified number of replicas.
Only works with Swarm or Compose modes. Single container
mode cannot be scaled (must stop and restart with --replicas).
Scaling is live and does not require downtime. The load
balancer will automatically distribute traffic to new replicas.
Examples:
# Scale up to 10 replicas
cnode scale 10
# Scale down to 2 replicas
cnode scale 2
# Scale to 1 (minimum)
cnode scale 1
"""
if replicas < 1:
console.print("[red]Error: Replicas must be at least 1[/red]")
return
manager = ServerManager()
with console.status(f"[cyan]Scaling to {replicas} replicas..."):
async def _scale():
return await manager.scale(replicas=replicas)
result = anyio.run(_scale)
if result["success"]:
console.print(Panel(
f"[green]✓ Scaled successfully[/green]\n\n"
f"New replica count: [bold]{replicas}[/bold]\n"
f"Mode: [cyan]{result.get('mode')}[/cyan]",
title="Scaling Complete",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Scaling failed[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "single container" in error_msg.lower():
console.print("\n[yellow]Hint: For single container mode:[/yellow]")
console.print("[yellow] 1. cnode stop[/yellow]")
console.print(f"[yellow] 2. cnode start --replicas {replicas}[/yellow]")
@cli.command("logs")
@click.option(
"--follow", "-f",
is_flag=True,
help="Follow log output (like tail -f)"
)
@click.option(
"--tail",
type=int,
default=100,
help="Number of lines to show (default: 100)"
)
def logs_cmd(follow: bool, tail: int):
"""View server logs.
Shows logs from running containers/services. Use --follow
to stream logs in real-time.
Examples:
# Show last 100 lines
cnode logs
# Show last 500 lines
cnode logs --tail 500
# Follow logs in real-time
cnode logs --follow
# Combine options
cnode logs -f --tail 50
"""
manager = ServerManager()
async def _logs():
return await manager.logs(follow=follow, tail=tail)
output = anyio.run(_logs)
console.print(output)
@cli.command("cleanup")
@click.option(
"--force",
is_flag=True,
help="Force cleanup even if state file doesn't exist"
)
def cleanup_cmd(force: bool):
"""Force cleanup of all Crawl4AI Docker resources.
Stops and removes all containers, networks, and optionally volumes.
Useful when server is stuck or state is corrupted.
Examples:
# Clean up everything
cnode cleanup
# Force cleanup (ignore state file)
cnode cleanup --force
"""
manager = ServerManager()
console.print(Panel(
f"[yellow]⚠️ Cleaning up Crawl4AI Docker resources[/yellow]\n\n"
f"This will stop and remove:\n"
f"- All Crawl4AI containers\n"
f"- Nginx load balancer\n"
f"- Redis instance\n"
f"- Docker networks\n"
f"- State files",
title="Cleanup",
border_style="yellow"
))
if not force and not Confirm.ask("[yellow]Continue with cleanup?[/yellow]"):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Cleaning up resources..."):
async def _cleanup():
return await manager.cleanup(force=force)
result = anyio.run(_cleanup)
if result["success"]:
console.print(Panel(
f"[green]✓ Cleanup completed successfully[/green]\n\n"
f"Removed: {result.get('removed', 0)} containers\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Cleanup Complete",
border_style="green"
))
else:
console.print(Panel(
f"[yellow]⚠️ Partial cleanup[/yellow]\n\n"
f"{result.get('message', 'Some resources may still exist')}",
title="Cleanup Status",
border_style="yellow"
))
@cli.command("restart")
@click.option(
"--replicas", "-r",
type=int,
help="New replica count (optional)"
)
def restart_cmd(replicas: int):
"""Restart server (stop then start with same config).
Preserves existing configuration unless overridden with options.
Useful for applying image updates or recovering from errors.
Examples:
# Restart with same configuration
cnode restart
# Restart and change replica count
cnode restart --replicas 5
"""
manager = ServerManager()
# Get current state
async def _get_status():
return await manager.status()
current = anyio.run(_get_status)
if not current["running"]:
console.print("[yellow]No server is running. Use 'cnode start' instead.[/yellow]")
return
# Extract current config
current_replicas = current.get("replicas", 1)
current_port = current.get("port", 11235)
current_image = current.get("image", "unclecode/crawl4ai:latest")
current_mode = current.get("mode", "auto")
# Override with CLI args
new_replicas = replicas if replicas is not None else current_replicas
console.print(Panel(
f"[cyan]Restarting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{current_replicas}[/yellow] → [green]{new_replicas}[/green]\n"
f"Port: [yellow]{current_port}[/yellow]\n"
f"Mode: [yellow]{current_mode}[/yellow]",
title="Server Restart",
border_style="cyan"
))
# Stop current
with console.status("[cyan]Stopping current server..."):
async def _stop_server():
return await manager.stop(remove_volumes=False)
stop_result = anyio.run(_stop_server)
if not stop_result["success"]:
console.print(f"[red]Failed to stop server: {stop_result.get('error')}[/red]")
return
# Start new
with console.status("[cyan]Starting server..."):
async def _start_server():
return await manager.start(
replicas=new_replicas,
mode="auto",
port=current_port,
image=current_image
)
start_result = anyio.run(_start_server)
if start_result["success"]:
console.print(Panel(
f"[green]✓ Server restarted successfully![/green]\n\n"
f"URL: [bold]http://localhost:{current_port}[/bold]",
title="Restart Complete",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Failed to restart server[/red]\n\n"
f"{start_result.get('error', 'Unknown error')}",
title="Error",
border_style="red"
))
def main():
"""Entry point for cnode CLI"""
cli()
if __name__ == "__main__":
main()
# Test comment

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

View File

@@ -5,7 +5,6 @@ import asyncio
from typing import Dict, List, Optional
from datetime import datetime, timezone
from collections import deque
from dataclasses import dataclass
from redis import asyncio as aioredis
from utils import get_container_memory_percent
import psutil
@@ -13,49 +12,13 @@ import logging
logger = logging.getLogger(__name__)
# ========== Configuration ==========
@dataclass
class RedisTTLConfig:
"""Redis TTL configuration (in seconds).
Configures how long different types of monitoring data are retained in Redis.
Adjust based on your monitoring needs and Redis memory constraints.
"""
active_requests: int = 300 # 5 minutes - short-lived active request data
completed_requests: int = 3600 # 1 hour - recent completed requests
janitor_events: int = 3600 # 1 hour - browser cleanup events
errors: int = 3600 # 1 hour - error logs
endpoint_stats: int = 86400 # 24 hours - aggregated endpoint statistics
heartbeat: int = 60 # 1 minute - container heartbeat (2x the 30s interval)
@classmethod
def from_env(cls) -> 'RedisTTLConfig':
"""Load TTL configuration from environment variables."""
import os
return cls(
active_requests=int(os.getenv('REDIS_TTL_ACTIVE_REQUESTS', 300)),
completed_requests=int(os.getenv('REDIS_TTL_COMPLETED_REQUESTS', 3600)),
janitor_events=int(os.getenv('REDIS_TTL_JANITOR_EVENTS', 3600)),
errors=int(os.getenv('REDIS_TTL_ERRORS', 3600)),
endpoint_stats=int(os.getenv('REDIS_TTL_ENDPOINT_STATS', 86400)),
heartbeat=int(os.getenv('REDIS_TTL_HEARTBEAT', 60)),
)
class MonitorStats:
"""Tracks real-time server stats with Redis persistence."""
def __init__(self, redis: aioredis.Redis, ttl_config: Optional[RedisTTLConfig] = None):
def __init__(self, redis: aioredis.Redis):
self.redis = redis
self.ttl = ttl_config or RedisTTLConfig.from_env()
self.start_time = time.time()
# Get container ID for Redis keys
from utils import get_container_id
self.container_id = get_container_id()
# In-memory queues (fast reads, Redis backup)
self.active_requests: Dict[str, Dict] = {} # id -> request info
self.completed_requests: deque = deque(maxlen=100) # Last 100
@@ -69,9 +32,6 @@ class MonitorStats:
self._persist_queue: asyncio.Queue = asyncio.Queue(maxsize=10)
self._persist_worker_task: Optional[asyncio.Task] = None
# Heartbeat task for container discovery
self._heartbeat_task: Optional[asyncio.Task] = None
# Timeline data (5min window, 5s resolution = 60 points)
self.memory_timeline: deque = deque(maxlen=60)
self.requests_timeline: deque = deque(maxlen=60)
@@ -85,14 +45,10 @@ class MonitorStats:
"url": url[:100], # Truncate long URLs
"start_time": time.time(),
"config_sig": config.get("sig", "default") if config else "default",
"mem_start": psutil.Process().memory_info().rss / (1024 * 1024),
"container_id": self.container_id
"mem_start": psutil.Process().memory_info().rss / (1024 * 1024)
}
self.active_requests[request_id] = req_info
# Persist to Redis
await self._persist_active_requests()
# Increment endpoint counter
if endpoint not in self.endpoint_stats:
self.endpoint_stats[endpoint] = {
@@ -139,29 +95,19 @@ class MonitorStats:
"success": success,
"error": error,
"status_code": status_code,
"pool_hit": pool_hit,
"container_id": self.container_id
"pool_hit": pool_hit
}
self.completed_requests.append(completed)
# Persist to Redis
await self._persist_completed_requests()
await self._persist_active_requests() # Update active (removed this request)
# Track errors
if not success and error:
error_entry = {
self.errors.append({
"timestamp": end_time,
"endpoint": endpoint,
"url": req_info["url"],
"error": error,
"request_id": request_id,
"message": error,
"level": "ERROR",
"container_id": self.container_id
}
self.errors.append(error_entry)
await self._persist_errors()
"request_id": request_id
})
await self._persist_endpoint_stats()
@@ -171,10 +117,8 @@ class MonitorStats:
"timestamp": time.time(),
"type": event_type, # "close_cold", "close_hot", "promote"
"sig": sig[:8],
"details": details,
"container_id": self.container_id
"details": details
})
await self._persist_janitor_events()
def _cleanup_old_entries(self, max_age_seconds: int = 300):
"""Remove entries older than max_age_seconds (default 5min)."""
@@ -205,23 +149,13 @@ class MonitorStats:
recent_reqs = sum(1 for req in self.completed_requests
if now - req.get("end_time", 0) < 5)
# Browser counts (acquire lock with timeout to prevent deadlock)
# Browser counts (acquire lock to prevent race conditions)
from crawler_pool import PERMANENT, HOT_POOL, COLD_POOL, LOCK
try:
async with asyncio.timeout(2.0):
async with LOCK:
browser_count = {
"permanent": 1 if PERMANENT else 0,
"hot": len(HOT_POOL),
"cold": len(COLD_POOL)
}
except asyncio.TimeoutError:
logger.warning("Lock acquisition timeout in update_timeline, using cached browser counts")
# Use last known values or defaults
async with LOCK:
browser_count = {
"permanent": 1,
"hot": 0,
"cold": 0
"permanent": 1 if PERMANENT else 0,
"hot": len(HOT_POOL),
"cold": len(COLD_POOL)
}
self.memory_timeline.append({"time": now, "value": mem_pct})
@@ -229,117 +163,15 @@ class MonitorStats:
self.browser_timeline.append({"time": now, "browsers": browser_count})
async def _persist_endpoint_stats(self):
"""Persist endpoint stats to Redis with retry logic."""
max_retries = 3
for attempt in range(max_retries):
try:
await self.redis.set(
"monitor:endpoint_stats",
json.dumps(self.endpoint_stats),
ex=self.ttl.endpoint_stats
)
return # Success
except aioredis.ConnectionError as e:
if attempt < max_retries - 1:
backoff = 0.5 * (2 ** attempt) # 0.5s, 1s, 2s
logger.warning(f"Redis connection error persisting endpoint stats (attempt {attempt + 1}/{max_retries}), retrying in {backoff}s: {e}")
await asyncio.sleep(backoff)
else:
logger.error(f"Failed to persist endpoint stats after {max_retries} attempts: {e}")
except Exception as e:
logger.error(f"Non-retryable error persisting endpoint stats: {e}")
break
async def _persist_active_requests(self):
"""Persist active requests to Redis with retry logic."""
max_retries = 3
for attempt in range(max_retries):
try:
if self.active_requests:
await self.redis.set(
f"monitor:{self.container_id}:active_requests",
json.dumps(list(self.active_requests.values())),
ex=self.ttl.active_requests
)
else:
await self.redis.delete(f"monitor:{self.container_id}:active_requests")
return # Success
except aioredis.ConnectionError as e:
if attempt < max_retries - 1:
backoff = 0.5 * (2 ** attempt) # 0.5s, 1s, 2s
logger.warning(f"Redis connection error persisting active requests (attempt {attempt + 1}/{max_retries}), retrying in {backoff}s: {e}")
await asyncio.sleep(backoff)
else:
logger.error(f"Failed to persist active requests after {max_retries} attempts: {e}")
except Exception as e:
logger.error(f"Non-retryable error persisting active requests: {e}")
break
async def _persist_completed_requests(self):
"""Persist completed requests to Redis with retry logic."""
max_retries = 3
for attempt in range(max_retries):
try:
await self.redis.set(
f"monitor:{self.container_id}:completed",
json.dumps(list(self.completed_requests)),
ex=self.ttl.completed_requests
)
return # Success
except aioredis.ConnectionError as e:
if attempt < max_retries - 1:
backoff = 0.5 * (2 ** attempt) # 0.5s, 1s, 2s
logger.warning(f"Redis connection error persisting completed requests (attempt {attempt + 1}/{max_retries}), retrying in {backoff}s: {e}")
await asyncio.sleep(backoff)
else:
logger.error(f"Failed to persist completed requests after {max_retries} attempts: {e}")
except Exception as e:
logger.error(f"Non-retryable error persisting completed requests: {e}")
break
async def _persist_janitor_events(self):
"""Persist janitor events to Redis with retry logic."""
max_retries = 3
for attempt in range(max_retries):
try:
await self.redis.set(
f"monitor:{self.container_id}:janitor",
json.dumps(list(self.janitor_events)),
ex=self.ttl.janitor_events
)
return # Success
except aioredis.ConnectionError as e:
if attempt < max_retries - 1:
backoff = 0.5 * (2 ** attempt) # 0.5s, 1s, 2s
logger.warning(f"Redis connection error persisting janitor events (attempt {attempt + 1}/{max_retries}), retrying in {backoff}s: {e}")
await asyncio.sleep(backoff)
else:
logger.error(f"Failed to persist janitor events after {max_retries} attempts: {e}")
except Exception as e:
logger.error(f"Non-retryable error persisting janitor events: {e}")
break
async def _persist_errors(self):
"""Persist errors to Redis with retry logic."""
max_retries = 3
for attempt in range(max_retries):
try:
await self.redis.set(
f"monitor:{self.container_id}:errors",
json.dumps(list(self.errors)),
ex=self.ttl.errors
)
return # Success
except aioredis.ConnectionError as e:
if attempt < max_retries - 1:
backoff = 0.5 * (2 ** attempt) # 0.5s, 1s, 2s
logger.warning(f"Redis connection error persisting errors (attempt {attempt + 1}/{max_retries}), retrying in {backoff}s: {e}")
await asyncio.sleep(backoff)
else:
logger.error(f"Failed to persist errors after {max_retries} attempts: {e}")
except Exception as e:
logger.error(f"Non-retryable error persisting errors: {e}")
break
"""Persist endpoint stats to Redis."""
try:
await self.redis.set(
"monitor:endpoint_stats",
json.dumps(self.endpoint_stats),
ex=86400 # 24h TTL
)
except Exception as e:
logger.warning(f"Failed to persist endpoint stats: {e}")
async def _persistence_worker(self):
"""Background worker to persist stats to Redis."""
@@ -370,121 +202,25 @@ class MonitorStats:
self._persist_worker_task = None
logger.info("Stopped persistence worker")
async def _heartbeat_worker(self):
"""Send heartbeat to Redis every 30s with circuit breaker for failures."""
from utils import detect_deployment_mode
import os
heartbeat_failures = 0
max_failures = 5 # Circuit breaker threshold
while True:
try:
# Get hostname/container name for friendly display
# Try HOSTNAME env var first (set by Docker Compose), then socket.gethostname()
import socket
hostname = os.getenv("HOSTNAME", socket.gethostname())
# Register this container
mode, containers = detect_deployment_mode()
container_info = {
"id": self.container_id,
"hostname": hostname,
"last_seen": time.time(),
"mode": mode,
"failure_count": heartbeat_failures
}
# Set heartbeat with configured TTL
await self.redis.setex(
f"monitor:heartbeat:{self.container_id}",
self.ttl.heartbeat,
json.dumps(container_info)
)
# Add to active containers set
await self.redis.sadd("monitor:active_containers", self.container_id)
# Reset failure counter on success
heartbeat_failures = 0
# Wait 30s before next heartbeat
await asyncio.sleep(30)
except asyncio.CancelledError:
break
except aioredis.ConnectionError as e:
heartbeat_failures += 1
logger.error(
f"Heartbeat Redis connection error (attempt {heartbeat_failures}/{max_failures}): {e}"
)
if heartbeat_failures >= max_failures:
# Circuit breaker - back off for longer
logger.critical(
f"Heartbeat circuit breaker triggered after {heartbeat_failures} failures. "
f"Container will appear offline for 5 minutes."
)
await asyncio.sleep(300) # 5 min backoff
heartbeat_failures = 0
else:
# Exponential backoff
backoff = min(30 * (2 ** heartbeat_failures), 300)
await asyncio.sleep(backoff)
except Exception as e:
logger.error(f"Unexpected heartbeat error: {e}", exc_info=True)
await asyncio.sleep(30)
def start_heartbeat(self):
"""Start the heartbeat worker."""
if not self._heartbeat_task:
self._heartbeat_task = asyncio.create_task(self._heartbeat_worker())
logger.info("Started heartbeat worker")
async def stop_heartbeat(self):
"""Stop the heartbeat worker and immediately deregister container."""
if self._heartbeat_task:
self._heartbeat_task.cancel()
try:
await self._heartbeat_task
except asyncio.CancelledError:
pass
# Immediate deregistration (no 60s wait)
try:
await self.redis.srem("monitor:active_containers", self.container_id)
await self.redis.delete(f"monitor:heartbeat:{self.container_id}")
logger.info(f"Container {self.container_id} immediately deregistered from monitoring")
except Exception as e:
logger.warning(f"Failed to deregister container on shutdown: {e}")
self._heartbeat_task = None
logger.info("Stopped heartbeat worker")
async def cleanup(self):
"""Cleanup on shutdown - persist final stats and stop workers."""
logger.info("Monitor cleanup starting...")
try:
# Persist final stats before shutdown
await self._persist_endpoint_stats()
# Stop background workers
# Stop background worker
await self.stop_persistence_worker()
await self.stop_heartbeat()
logger.info("Monitor cleanup completed")
except Exception as e:
logger.error(f"Monitor cleanup error: {e}")
async def load_from_redis(self):
"""Load persisted stats from Redis and start workers."""
"""Load persisted stats from Redis."""
try:
data = await self.redis.get("monitor:endpoint_stats")
if data:
self.endpoint_stats = json.loads(data)
logger.info("Loaded endpoint stats from Redis")
# Start background workers
self.start_heartbeat()
except Exception as e:
logger.warning(f"Failed to load from Redis: {e}")
@@ -496,28 +232,17 @@ class MonitorStats:
# Network I/O (delta since last call)
net = psutil.net_io_counters()
# Pool status (acquire lock with timeout to prevent race conditions)
# Pool status (acquire lock to prevent race conditions)
from crawler_pool import PERMANENT, HOT_POOL, COLD_POOL, LOCK
try:
async with asyncio.timeout(2.0):
async with LOCK:
# TODO: Track actual browser process memory instead of estimates
# These are conservative estimates based on typical Chromium usage
permanent_mem = 270 if PERMANENT else 0 # Estimate: ~270MB for permanent browser
hot_mem = len(HOT_POOL) * 180 # Estimate: ~180MB per hot pool browser
cold_mem = len(COLD_POOL) * 180 # Estimate: ~180MB per cold pool browser
permanent_active = PERMANENT is not None
hot_count = len(HOT_POOL)
cold_count = len(COLD_POOL)
except asyncio.TimeoutError:
logger.warning("Lock acquisition timeout in get_health_summary, using defaults")
# Use safe defaults when lock times out
permanent_mem = 0
hot_mem = 0
cold_mem = 0
permanent_active = False
hot_count = 0
cold_count = 0
async with LOCK:
# TODO: Track actual browser process memory instead of estimates
# These are conservative estimates based on typical Chromium usage
permanent_mem = 270 if PERMANENT else 0 # Estimate: ~270MB for permanent browser
hot_mem = len(HOT_POOL) * 180 # Estimate: ~180MB per hot pool browser
cold_mem = len(COLD_POOL) * 180 # Estimate: ~180MB per cold pool browser
permanent_active = PERMANENT is not None
hot_count = len(HOT_POOL)
cold_count = len(COLD_POOL)
return {
"container": {
@@ -561,52 +286,46 @@ class MonitorStats:
return requests
async def get_browser_list(self) -> List[Dict]:
"""Get detailed browser pool information with timeout protection."""
"""Get detailed browser pool information."""
from crawler_pool import PERMANENT, HOT_POOL, COLD_POOL, LAST_USED, USAGE_COUNT, DEFAULT_CONFIG_SIG, LOCK
browsers = []
now = time.time()
# Acquire lock with timeout to prevent deadlock
try:
async with asyncio.timeout(2.0):
async with LOCK:
if PERMANENT:
browsers.append({
"type": "permanent",
"sig": DEFAULT_CONFIG_SIG[:8] if DEFAULT_CONFIG_SIG else "unknown",
"age_seconds": int(now - self.start_time),
"last_used_seconds": int(now - LAST_USED.get(DEFAULT_CONFIG_SIG, now)),
"memory_mb": 270,
"hits": USAGE_COUNT.get(DEFAULT_CONFIG_SIG, 0),
"killable": False
})
# Acquire lock to prevent race conditions during iteration
async with LOCK:
if PERMANENT:
browsers.append({
"type": "permanent",
"sig": DEFAULT_CONFIG_SIG[:8] if DEFAULT_CONFIG_SIG else "unknown",
"age_seconds": int(now - self.start_time),
"last_used_seconds": int(now - LAST_USED.get(DEFAULT_CONFIG_SIG, now)),
"memory_mb": 270,
"hits": USAGE_COUNT.get(DEFAULT_CONFIG_SIG, 0),
"killable": False
})
for sig, crawler in HOT_POOL.items():
browsers.append({
"type": "hot",
"sig": sig[:8],
"age_seconds": int(now - self.start_time), # Approximation
"last_used_seconds": int(now - LAST_USED.get(sig, now)),
"memory_mb": 180, # Estimate
"hits": USAGE_COUNT.get(sig, 0),
"killable": True
})
for sig, crawler in HOT_POOL.items():
browsers.append({
"type": "hot",
"sig": sig[:8],
"age_seconds": int(now - self.start_time), # Approximation
"last_used_seconds": int(now - LAST_USED.get(sig, now)),
"memory_mb": 180, # Estimate
"hits": USAGE_COUNT.get(sig, 0),
"killable": True
})
for sig, crawler in COLD_POOL.items():
browsers.append({
"type": "cold",
"sig": sig[:8],
"age_seconds": int(now - self.start_time),
"last_used_seconds": int(now - LAST_USED.get(sig, now)),
"memory_mb": 180,
"hits": USAGE_COUNT.get(sig, 0),
"killable": True
})
except asyncio.TimeoutError:
logger.error("Browser list lock timeout - pool may be locked by janitor")
# Return empty list when lock times out to prevent blocking
return []
for sig, crawler in COLD_POOL.items():
browsers.append({
"type": "cold",
"sig": sig[:8],
"age_seconds": int(now - self.start_time),
"last_used_seconds": int(now - LAST_USED.get(sig, now)),
"memory_mb": 180,
"hits": USAGE_COUNT.get(sig, 0),
"killable": True
})
return browsers

View File

@@ -3,140 +3,14 @@ from fastapi import APIRouter, HTTPException, WebSocket, WebSocketDisconnect
from pydantic import BaseModel
from typing import Optional
from monitor import get_monitor
from utils import detect_deployment_mode, get_container_id
import logging
import asyncio
import json
import re
logger = logging.getLogger(__name__)
router = APIRouter(prefix="/monitor", tags=["monitor"])
# ========== Security & Validation ==========
def validate_container_id(cid: str) -> bool:
"""Validate container ID format to prevent Redis key injection.
Docker container IDs are 12-64 character hexadecimal strings.
Hostnames are alphanumeric with dashes and underscores.
Args:
cid: Container ID to validate
Returns:
True if valid, False otherwise
"""
if not cid or not isinstance(cid, str):
return False
# Allow alphanumeric, dashes, and underscores only (1-64 chars)
# This prevents path traversal (../../), wildcards (**), and other injection attempts
return bool(re.match(r'^[a-zA-Z0-9_-]{1,64}$', cid))
# ========== Redis Aggregation Helpers ==========
async def _get_active_containers():
"""Get list of active container IDs from Redis with validation."""
try:
monitor = get_monitor()
container_ids = await monitor.redis.smembers("monitor:active_containers")
# Decode and validate each container ID
validated = []
for cid in container_ids:
cid_str = cid.decode() if isinstance(cid, bytes) else cid
if validate_container_id(cid_str):
validated.append(cid_str)
else:
logger.warning(f"Invalid container ID format rejected: {cid_str}")
return validated
except Exception as e:
logger.error(f"Failed to get active containers: {e}")
return []
async def _aggregate_active_requests():
"""Aggregate active requests from all containers."""
container_ids = await _get_active_containers()
all_requests = []
monitor = get_monitor()
for container_id in container_ids:
try:
data = await monitor.redis.get(f"monitor:{container_id}:active_requests")
if data:
requests = json.loads(data)
all_requests.extend(requests)
except Exception as e:
logger.warning(f"Failed to get active requests from {container_id}: {e}")
return all_requests
async def _aggregate_completed_requests(limit=100):
"""Aggregate completed requests from all containers."""
container_ids = await _get_active_containers()
all_requests = []
monitor = get_monitor()
for container_id in container_ids:
try:
data = await monitor.redis.get(f"monitor:{container_id}:completed")
if data:
requests = json.loads(data)
all_requests.extend(requests)
except Exception as e:
logger.warning(f"Failed to get completed requests from {container_id}: {e}")
# Sort by end_time (most recent first) and limit
all_requests.sort(key=lambda x: x.get("end_time", 0), reverse=True)
return all_requests[:limit]
async def _aggregate_janitor_events(limit=100):
"""Aggregate janitor events from all containers."""
container_ids = await _get_active_containers()
all_events = []
monitor = get_monitor()
for container_id in container_ids:
try:
data = await monitor.redis.get(f"monitor:{container_id}:janitor")
if data:
events = json.loads(data)
all_events.extend(events)
except Exception as e:
logger.warning(f"Failed to get janitor events from {container_id}: {e}")
# Sort by timestamp (most recent first) and limit
all_events.sort(key=lambda x: x.get("timestamp", 0), reverse=True)
return all_events[:limit]
async def _aggregate_errors(limit=100):
"""Aggregate errors from all containers."""
container_ids = await _get_active_containers()
all_errors = []
monitor = get_monitor()
for container_id in container_ids:
try:
data = await monitor.redis.get(f"monitor:{container_id}:errors")
if data:
errors = json.loads(data)
all_errors.extend(errors)
except Exception as e:
logger.warning(f"Failed to get errors from {container_id}: {e}")
# Sort by timestamp (most recent first) and limit
all_errors.sort(key=lambda x: x.get("timestamp", 0), reverse=True)
return all_errors[:limit]
@router.get("/health")
async def get_health():
"""Get current system health snapshot."""
@@ -163,23 +37,18 @@ async def get_requests(status: str = "all", limit: int = 50):
raise HTTPException(400, f"Invalid limit: {limit}. Must be between 1 and 1000")
try:
# Aggregate from all containers via Redis
active_requests = await _aggregate_active_requests()
completed_requests = await _aggregate_completed_requests(limit)
# Filter by status if needed
if status in ["success", "error"]:
is_success = (status == "success")
completed_requests = [r for r in completed_requests if r.get("success") == is_success]
monitor = get_monitor()
if status == "active":
return {"active": active_requests, "completed": []}
return {"active": monitor.get_active_requests(), "completed": []}
elif status == "completed":
return {"active": [], "completed": completed_requests}
else: # "all" or success/error
return {"active": [], "completed": monitor.get_completed_requests(limit)}
elif status in ["success", "error"]:
return {"active": [], "completed": monitor.get_completed_requests(limit, status)}
else: # "all"
return {
"active": active_requests,
"completed": completed_requests
"active": monitor.get_active_requests(),
"completed": monitor.get_completed_requests(limit)
}
except Exception as e:
logger.error(f"Error getting requests: {e}")
@@ -191,13 +60,8 @@ async def get_browsers():
"""Get detailed browser pool information."""
try:
monitor = get_monitor()
container_id = get_container_id()
browsers = await monitor.get_browser_list()
# Add container_id to each browser
for browser in browsers:
browser["container_id"] = container_id
# Calculate summary stats
total_browsers = len(browsers)
total_memory = sum(b["memory_mb"] for b in browsers)
@@ -213,8 +77,7 @@ async def get_browsers():
"total_count": total_browsers,
"total_memory_mb": total_memory,
"reuse_rate_percent": round(reuse_rate, 1)
},
"container_id": container_id
}
}
except Exception as e:
logger.error(f"Error getting browsers: {e}")
@@ -262,9 +125,8 @@ async def get_janitor_log(limit: int = 100):
raise HTTPException(400, f"Invalid limit: {limit}. Must be between 1 and 1000")
try:
# Aggregate from all containers via Redis
events = await _aggregate_janitor_events(limit)
return {"events": events}
monitor = get_monitor()
return {"events": monitor.get_janitor_log(limit)}
except Exception as e:
logger.error(f"Error getting janitor log: {e}")
raise HTTPException(500, str(e))
@@ -278,9 +140,8 @@ async def get_errors_log(limit: int = 100):
raise HTTPException(400, f"Invalid limit: {limit}. Must be between 1 and 1000")
try:
# Aggregate from all containers via Redis
errors = await _aggregate_errors(limit)
return {"errors": errors}
monitor = get_monitor()
return {"errors": monitor.get_errors_log(limit)}
except Exception as e:
logger.error(f"Error getting errors log: {e}")
raise HTTPException(500, str(e))
@@ -489,57 +350,15 @@ async def reset_stats():
raise HTTPException(500, str(e))
@router.get("/containers")
async def get_containers():
"""Get container deployment info from Redis heartbeats."""
try:
monitor = get_monitor()
container_ids = await _get_active_containers()
containers = []
for cid in container_ids:
try:
# Get heartbeat data
data = await monitor.redis.get(f"monitor:heartbeat:{cid}")
if data:
info = json.loads(data)
containers.append({
"id": info.get("id", cid),
"hostname": info.get("hostname", cid),
"healthy": True # If heartbeat exists, it's healthy
})
except Exception as e:
logger.warning(f"Failed to get heartbeat for {cid}: {e}")
# Determine mode
mode = "single" if len(containers) == 1 else "compose"
if len(containers) > 1:
# Check if any hostname has swarm pattern (service.slot.task_id)
if any("." in c["hostname"] and len(c["hostname"].split(".")) > 2 for c in containers):
mode = "swarm"
return {
"mode": mode,
"container_id": get_container_id(),
"containers": containers,
"count": len(containers)
}
except Exception as e:
logger.error(f"Error getting containers: {e}")
raise HTTPException(500, str(e))
@router.websocket("/ws")
async def websocket_endpoint(websocket: WebSocket):
"""WebSocket endpoint for real-time monitoring updates.
Sends aggregated updates every 2 seconds from all containers with:
- Health stats (local container)
- Active/completed requests (aggregated from all containers)
- Browser pool status (local container only - not in Redis)
- Timeline data (local container - TODO: aggregate from Redis)
- Janitor events (aggregated from all containers)
- Errors (aggregated from all containers)
Sends updates every 2 seconds with:
- Health stats
- Active/completed requests
- Browser pool status
- Timeline data
"""
await websocket.accept()
logger.info("WebSocket client connected")
@@ -547,46 +366,24 @@ async def websocket_endpoint(websocket: WebSocket):
try:
while True:
try:
# Gather aggregated monitoring data from Redis
# Gather all monitoring data
monitor = get_monitor()
container_id = get_container_id()
# Get container info
containers_info = await get_containers()
# AGGREGATE data from all containers via Redis
active_reqs = await _aggregate_active_requests()
completed_reqs = await _aggregate_completed_requests(limit=10)
janitor_events = await _aggregate_janitor_events(limit=10)
errors_log = await _aggregate_errors(limit=10)
# Local container data (not aggregated)
local_health = await monitor.get_health_summary()
browsers = await monitor.get_browser_list() # Browser list is local only
# Add container_id to browsers (they're local)
for browser in browsers:
browser["container_id"] = container_id
data = {
"timestamp": asyncio.get_event_loop().time(),
"container_id": container_id, # This container handling the WebSocket
"is_aggregated": True, # Flag to indicate aggregated data
"local_health": local_health, # This container's health
"containers": containers_info.get("containers", []), # All containers
"health": await monitor.get_health_summary(),
"requests": {
"active": active_reqs, # Aggregated from all containers
"completed": completed_reqs # Aggregated from all containers
"active": monitor.get_active_requests(),
"completed": monitor.get_completed_requests(limit=10)
},
"browsers": browsers, # Local only (not in Redis)
"browsers": await monitor.get_browser_list(),
"timeline": {
# TODO: Aggregate timeline from Redis (currently local only)
"memory": monitor.get_timeline_data("memory", "5m"),
"requests": monitor.get_timeline_data("requests", "5m"),
"browsers": monitor.get_timeline_data("browsers", "5m")
},
"janitor": janitor_events, # Aggregated from all containers
"errors": errors_log # Aggregated from all containers
"janitor": monitor.get_janitor_log(limit=10),
"errors": monitor.get_errors_log(limit=10)
}
# Send update to client

View File

@@ -200,11 +200,7 @@ async def root():
return RedirectResponse("/playground")
# ─────────────────── infra / middleware ─────────────────────
# Build Redis URL from environment or config
redis_host = os.getenv("REDIS_HOST", config["redis"].get("host", "localhost"))
redis_port = os.getenv("REDIS_PORT", config["redis"].get("port", 6379))
redis_url = config["redis"].get("uri") or f"redis://{redis_host}:{redis_port}"
redis = aioredis.from_url(redis_url)
redis = aioredis.from_url(config["redis"].get("uri", "redis://localhost"))
limiter = Limiter(
key_func=get_remote_address,

File diff suppressed because it is too large Load Diff

View File

@@ -116,107 +116,74 @@
<!-- Main Content -->
<main class="flex-1 overflow-auto p-4 space-y-4">
<!-- System Health & Infrastructure (side by side) -->
<div class="grid grid-cols-2 gap-4">
<!-- System Health -->
<section class="bg-surface rounded-lg border border-border p-3">
<h2 class="text-sm font-medium mb-2 text-primary">System Health</h2>
<!-- System Health Bar -->
<section class="bg-surface rounded-lg border border-border p-4">
<h2 class="text-sm font-medium mb-3 text-primary">System Health</h2>
<!-- Row 1: CPU and Memory -->
<div class="grid grid-cols-2 gap-3 mb-2">
<!-- CPU -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">CPU</span>
<span id="cpu-percent" class="text-light">--%</span>
</div>
<div class="w-full bg-dark rounded-full h-2">
<div id="cpu-bar" class="progress-bar h-2 rounded-full bg-primary" style="width: 0%"></div>
</div>
<div class="grid grid-cols-4 gap-4 mb-4">
<!-- CPU -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">CPU</span>
<span id="cpu-percent" class="text-light">--%</span>
</div>
<!-- Memory -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Memory</span>
<span id="mem-percent" class="text-light">--%</span>
</div>
<div class="w-full bg-dark rounded-full h-2">
<div id="mem-bar" class="progress-bar h-2 rounded-full bg-accent" style="width: 0%"></div>
</div>
<div class="w-full bg-dark rounded-full h-2">
<div id="cpu-bar" class="progress-bar h-2 rounded-full bg-primary" style="width: 0%"></div>
</div>
</div>
<!-- Row 2: Network and Uptime -->
<div class="grid grid-cols-2 gap-3 mb-2">
<!-- Network -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Network</span>
<span id="net-io" class="text-light">--</span>
</div>
<div class="text-xs text-secondary"><span id="net-sent">0</span> / ⬇<span id="net-recv">0</span> MB</div>
<!-- Memory -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Memory</span>
<span id="mem-percent" class="text-light">--%</span>
</div>
<!-- Uptime -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Uptime</span>
<span id="uptime" class="text-light">--</span>
</div>
<div class="text-xs text-secondary" id="last-update">Live: --:--:--</div>
<div class="w-full bg-dark rounded-full h-2">
<div id="mem-bar" class="progress-bar h-2 rounded-full bg-accent" style="width: 0%"></div>
</div>
</div>
<!-- Pool Status -->
<div class="border-t border-border pt-2">
<div class="grid grid-cols-3 gap-3 text-xs">
<div>
<span class="text-secondary">🔥 Permanent:</span>
<span id="pool-perm" class="text-primary ml-1">INACTIVE (0MB)</span>
</div>
<div>
<span class="text-secondary">♨️ Hot:</span>
<span id="pool-hot" class="text-accent ml-1">0 (0MB)</span>
</div>
<div>
<span class="text-secondary">❄️ Cold:</span>
<span id="pool-cold" class="text-light ml-1">0 (0MB)</span>
</div>
</div>
<div class="mt-1 text-xs text-secondary">
<span>Janitor: </span><span id="janitor-status">adaptive</span> |
<span>Memory pressure: </span><span id="mem-pressure">LOW</span>
<!-- Network -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Network</span>
<span id="net-io" class="text-light">--</span>
</div>
<div class="text-xs text-secondary"><span id="net-sent">0</span> MB / ⬇<span id="net-recv">0</span> MB</div>
</div>
</section>
<!-- Infrastructure Section -->
<section id="containers-section" class="bg-surface rounded-lg border border-border p-3" style="display: none;">
<div class="flex items-center justify-between mb-3">
<h2 class="text-sm font-medium text-primary">📦 Infrastructure</h2>
<div class="flex items-center space-x-2">
<span class="text-xs text-secondary">Mode:</span>
<span id="deployment-mode" class="text-xs text-primary font-medium">single</span>
<span class="text-xs text-secondary">|</span>
<span class="text-xs text-secondary">Containers:</span>
<span id="container-count" class="text-xs text-accent font-medium">1</span>
<!-- Uptime -->
<div>
<div class="flex justify-between text-xs mb-1">
<span class="text-secondary">Uptime</span>
<span id="uptime" class="text-light">--</span>
</div>
<div class="text-xs text-secondary" id="last-update">Updated: never</div>
</div>
</div>
<!-- Container Filter Buttons -->
<div id="container-filters" class="flex flex-wrap gap-2 mb-3">
<button class="container-filter-btn px-3 py-1 rounded text-xs bg-primary text-dark font-medium" data-container="all">
All
</button>
</div>
<!-- Container Grid -->
<div id="containers-grid" class="grid grid-cols-3 gap-3 text-xs">
<!-- Containers will be populated here -->
<!-- Pool Status -->
<div class="border-t border-border pt-3">
<div class="grid grid-cols-3 gap-4 text-xs">
<div>
<span class="text-secondary">🔥 Permanent:</span>
<span id="pool-perm" class="text-primary ml-2">INACTIVE (0MB)</span>
</div>
<div>
<span class="text-secondary">♨️ Hot:</span>
<span id="pool-hot" class="text-accent ml-2">0 (0MB)</span>
</div>
<div>
<span class="text-secondary">❄️ Cold:</span>
<span id="pool-cold" class="text-light ml-2">0 (0MB)</span>
</div>
</div>
<div class="mt-2 text-xs text-secondary">
<span>Janitor: </span><span id="janitor-status">adaptive</span> |
<span>Memory pressure: </span><span id="mem-pressure">LOW</span>
</div>
</div>
</section>
</div>
<!-- Live Activity Grid (2x2) -->
<div class="grid grid-cols-2 gap-4">
@@ -256,12 +223,11 @@
<th class="py-1 pr-2">Age</th>
<th class="py-1 pr-2">Used</th>
<th class="py-1 pr-2">Hits</th>
<th class="py-1 pr-2">Container</th>
<th class="py-1">Act</th>
</tr>
</thead>
<tbody id="browsers-table-body">
<tr><td colspan="7" class="text-center py-4 text-secondary">No browsers</td></tr>
<tr><td colspan="6" class="text-center py-4 text-secondary">No browsers</td></tr>
</tbody>
</table>
</div>
@@ -390,16 +356,6 @@
}
function connectWebSocket() {
// Clean up existing connection first to prevent resource leaks
if (websocket) {
try {
websocket.close();
} catch (e) {
console.error('Error closing old WebSocket:', e);
}
websocket = null;
}
if (wsReconnectAttempts >= MAX_WS_RECONNECT) {
console.log('Max WebSocket reconnect attempts reached, falling back to polling');
useWebSocket = false;
@@ -414,24 +370,9 @@
const protocol = window.location.protocol === 'https:' ? 'wss:' : 'ws:';
const wsUrl = `${protocol}//${window.location.host}/monitor/ws`;
try {
websocket = new WebSocket(wsUrl);
} catch (e) {
console.error('Failed to create WebSocket:', e);
setTimeout(() => connectWebSocket(), 2000 * wsReconnectAttempts);
return;
}
// Set connection timeout to prevent indefinite connection attempts
const connectionTimeout = setTimeout(() => {
if (websocket && websocket.readyState === WebSocket.CONNECTING) {
console.log('WebSocket connection timeout');
websocket.close();
}
}, 5000);
websocket = new WebSocket(wsUrl);
websocket.onopen = () => {
clearTimeout(connectionTimeout);
console.log('WebSocket connected');
wsReconnectAttempts = 0;
updateConnectionStatus('connected');
@@ -444,19 +385,15 @@
};
websocket.onerror = (error) => {
clearTimeout(connectionTimeout);
console.error('WebSocket error:', error);
};
websocket.onclose = (event) => {
clearTimeout(connectionTimeout);
console.log(`WebSocket closed: code=${event.code}, reason=${event.reason}`);
websocket.onclose = () => {
console.log('WebSocket closed');
updateConnectionStatus('disconnected', 'Reconnecting...');
websocket = null; // Clear reference
if (useWebSocket && wsReconnectAttempts < MAX_WS_RECONNECT) {
setTimeout(() => connectWebSocket(), 2000 * wsReconnectAttempts);
if (useWebSocket) {
setTimeout(connectWebSocket, 2000 * wsReconnectAttempts);
} else {
startAutoRefresh();
}
@@ -522,28 +459,18 @@
}
function updateRequestsDisplay(requests) {
// Filter requests based on current container filter
const filteredActive = currentContainerFilter === 'all'
? requests.active
: requests.active.filter(r => r.container_id === currentContainerFilter);
const filteredCompleted = currentContainerFilter === 'all'
? requests.completed
: requests.completed.filter(r => r.container_id === currentContainerFilter);
// Update active requests count
const activeCount = document.getElementById('active-count');
if (activeCount) activeCount.textContent = filteredActive.length;
if (activeCount) activeCount.textContent = requests.active.length;
// Update active requests list
const activeList = document.getElementById('active-requests-list');
if (activeList) {
if (filteredActive.length === 0) {
if (requests.active.length === 0) {
activeList.innerHTML = '<div class="text-secondary text-center py-2">No active requests</div>';
} else {
activeList.innerHTML = filteredActive.map(req => `
activeList.innerHTML = requests.active.map(req => `
<div class="flex items-center justify-between p-2 bg-dark rounded border border-border">
<span class="text-accent text-xs">${getContainerLabel(req.container_id)}</span>
<span class="text-primary">${req.id.substring(0, 8)}</span>
<span class="text-secondary">${req.endpoint}</span>
<span class="text-light truncate max-w-[200px]" title="${req.url}">${req.url}</span>
@@ -557,12 +484,11 @@
// Update completed requests
const completedList = document.getElementById('completed-requests-list');
if (completedList) {
if (filteredCompleted.length === 0) {
if (requests.completed.length === 0) {
completedList.innerHTML = '<div class="text-secondary text-center py-2">No completed requests</div>';
} else {
completedList.innerHTML = filteredCompleted.map(req => `
completedList.innerHTML = requests.completed.map(req => `
<div class="flex items-center gap-3 p-2 bg-dark rounded">
<span class="text-accent text-xs w-12 flex-shrink-0">${getContainerLabel(req.container_id)}</span>
<span class="text-secondary w-16 flex-shrink-0">${req.id.substring(0, 8)}</span>
<span class="text-secondary w-16 flex-shrink-0">${req.endpoint}</span>
<span class="text-light truncate flex-1" title="${req.url}">${req.url}</span>
@@ -585,14 +511,6 @@
const typeIcon = b.type === 'permanent' ? '🔥' : b.type === 'hot' ? '♨️' : '❄️';
const typeColor = b.type === 'permanent' ? 'text-primary' : b.type === 'hot' ? 'text-accent' : 'text-light';
// Check if should display based on filter
const shouldDisplay = currentContainerFilter === 'all' ||
b.container_id === currentContainerFilter;
if (!shouldDisplay) return '';
// Find container label (C-1, C-2, etc)
const containerLabel = getContainerLabel(b.container_id);
return `
<tr class="border-t border-border hover:bg-dark">
<td class="py-1 pr-2"><span class="${typeColor}">${typeIcon} ${b.type}</span></td>
@@ -600,7 +518,6 @@
<td class="py-1 pr-2">${formatSeconds(b.age_seconds || 0)}</td>
<td class="py-1 pr-2">${formatSeconds(b.last_used_seconds || 0)}</td>
<td class="py-1 pr-2">${b.hits}</td>
<td class="py-1 pr-2 text-accent text-xs">${containerLabel}</td>
<td class="py-1">
${b.killable ? `
<button onclick="killBrowser('${b.sig}')" class="text-red-500 hover:underline text-xs">X</button>
@@ -636,23 +553,16 @@
function updateJanitorDisplay(events) {
const janitorLog = document.getElementById('janitor-log');
if (janitorLog) {
// Filter events based on current container filter
const filtered = currentContainerFilter === 'all'
? events
: events.filter(e => e.container_id === currentContainerFilter);
if (filtered.length === 0) {
if (events.length === 0) {
janitorLog.innerHTML = '<div class="text-secondary text-center py-4">No events yet</div>';
} else {
janitorLog.innerHTML = filtered.slice(0, 10).reverse().map(evt => {
janitorLog.innerHTML = events.slice(0, 10).reverse().map(evt => {
const time = new Date(evt.timestamp * 1000).toLocaleTimeString();
const icon = evt.type === 'close_cold' ? '🧹❄️' : evt.type === 'close_hot' ? '🧹♨️' : '⬆️';
const details = JSON.stringify(evt.details);
const containerLabel = getContainerLabel(evt.container_id);
return `<div class="p-2 bg-dark rounded">
<span class="text-accent text-xs">${containerLabel}</span>
<span class="text-secondary ml-2">${time}</span>
<span class="text-secondary">${time}</span>
<span>${icon}</span>
<span class="text-primary">${evt.type}</span>
<span class="text-secondary">sig=${evt.sig}</span>
@@ -1149,90 +1059,10 @@
return `${m}m ${s}s`;
}
// ========== Containers Management ==========
let currentContainerFilter = 'all';
let containerMapping = {}; // Maps container_id to label (C-1, C-2, etc)
// Helper to get container label from ID or hostname
function getContainerLabel(containerId) {
// Try direct lookup first (works for both hostname and id)
if (containerMapping[containerId]) {
return containerMapping[containerId];
}
// Fallback: show first 8 chars of container ID
return containerId?.substring(0, 8) || 'unknown';
}
async function fetchContainers() {
try {
const res = await fetch('/monitor/containers');
const data = await res.json();
document.getElementById('deployment-mode').textContent = data.mode;
document.getElementById('container-count').textContent = data.count;
// Build container ID to label mapping
// Use hostname as primary key (friendly name like "crawl4ai-1")
// Also map id for backwards compatibility
containerMapping = {};
data.containers.forEach((c, i) => {
const label = `C-${i+1}`;
containerMapping[c.hostname] = label; // Map hostname
containerMapping[c.id] = label; // Also map id
});
// Show section only if multi-container
const section = document.getElementById('containers-section');
if (data.count > 1) {
section.style.display = 'block';
// Update filter buttons
const filtersDiv = document.getElementById('container-filters');
filtersDiv.innerHTML = `
<button class="container-filter-btn px-3 py-1 rounded text-xs ${currentContainerFilter === 'all' ? 'bg-primary text-dark' : 'bg-dark text-secondary'} font-medium" data-container="all">All</button>
${data.containers.map((c, i) => `
<button class="container-filter-btn px-3 py-1 rounded text-xs ${currentContainerFilter === c.id ? 'bg-primary text-dark' : 'bg-dark text-secondary'}" data-container="${c.id}">C-${i+1}</button>
`).join('')}
`;
// Add click handlers to filter buttons
document.querySelectorAll('.container-filter-btn').forEach(btn => {
btn.addEventListener('click', () => {
currentContainerFilter = btn.dataset.container;
fetchContainers(); // Refresh to update button styles
// Re-fetch all data with filter applied
fetchRequests();
fetchBrowsers();
fetchJanitorLogs();
fetchErrorLogs();
});
});
// Update containers grid
const grid = document.getElementById('containers-grid');
grid.innerHTML = data.containers.map((c, i) => `
<div class="p-3 bg-dark rounded border ${currentContainerFilter === c.id || currentContainerFilter === 'all' ? 'border-primary' : 'border-border'}">
<div class="flex items-center justify-between mb-2">
<span class="text-primary font-medium">C-${i+1}</span>
<span class="text-xs ${c.healthy ? 'text-accent' : 'text-red-500'}">${c.healthy ? '🟢' : '🔴'}</span>
</div>
<div class="text-xs text-secondary truncate" title="${c.hostname}">${c.hostname}</div>
</div>
`).join('');
} else {
section.style.display = 'none';
}
} catch (e) {
console.error('Failed to fetch containers:', e);
}
}
// ========== Filter change handler ==========
document.getElementById('filter-requests')?.addEventListener('change', fetchRequests);
// ========== Initialize ==========
// Fetch containers info on load
fetchContainers();
// Try WebSocket first, fallback to polling on failure
connectWebSocket();
</script>

View File

@@ -1,298 +0,0 @@
# Crawl4AI CLI E2E Test Suite
Comprehensive end-to-end tests for the `crwl server` command-line interface.
## Overview
This test suite validates all aspects of the Docker server CLI including:
- Basic operations (start, stop, status, logs)
- Advanced features (scaling, modes, custom configurations)
- Resource management and stress testing
- Dashboard UI functionality
- Edge cases and error handling
**Total Tests:** 32
- Basic: 8 tests
- Advanced: 8 tests
- Resource: 5 tests
- Dashboard: 1 test
- Edge Cases: 10 tests
## Prerequisites
```bash
# Activate virtual environment
source venv/bin/activate
# For dashboard tests, install Playwright
pip install playwright
playwright install chromium
# Ensure Docker is running
docker ps
```
## Quick Start
```bash
# Run all tests (except dashboard)
./run_tests.sh
# Run specific category
./run_tests.sh basic
./run_tests.sh advanced
./run_tests.sh resource
./run_tests.sh edge
# Run dashboard tests (slower, includes UI screenshots)
./run_tests.sh dashboard
# Run specific test
./run_tests.sh basic 01
./run_tests.sh edge 05
```
## Test Categories
### 1. Basic Tests (`basic/`)
Core CLI functionality tests.
| Test | Description | Expected Result |
|------|-------------|----------------|
| `test_01_start_default.sh` | Start server with defaults | 1 replica on port 11235 |
| `test_02_status.sh` | Check server status | Shows running state and details |
| `test_03_stop.sh` | Stop server | Clean shutdown, port freed |
| `test_04_start_custom_port.sh` | Start on port 8080 | Server on custom port |
| `test_05_start_replicas.sh` | Start with 3 replicas | Multi-container deployment |
| `test_06_logs.sh` | View server logs | Logs displayed correctly |
| `test_07_restart.sh` | Restart server | Preserves configuration |
| `test_08_cleanup.sh` | Force cleanup | All resources removed |
### 2. Advanced Tests (`advanced/`)
Advanced features and configurations.
| Test | Description | Expected Result |
|------|-------------|----------------|
| `test_01_scale_up.sh` | Scale 3 → 5 replicas | Live scaling without downtime |
| `test_02_scale_down.sh` | Scale 5 → 2 replicas | Graceful container removal |
| `test_03_mode_single.sh` | Explicit single mode | Single container deployment |
| `test_04_mode_compose.sh` | Compose mode with Nginx | Multi-container with load balancer |
| `test_05_custom_image.sh` | Custom image specification | Uses specified image tag |
| `test_06_env_file.sh` | Environment file loading | Variables loaded correctly |
| `test_07_stop_remove_volumes.sh` | Stop with volume removal | Volumes cleaned up |
| `test_08_restart_with_scale.sh` | Restart with new replica count | Configuration updated |
### 3. Resource Tests (`resource/`)
Resource monitoring and stress testing.
| Test | Description | Expected Result |
|------|-------------|----------------|
| `test_01_memory_monitoring.sh` | Monitor memory usage | Stats accessible and reasonable |
| `test_02_cpu_stress.sh` | Concurrent request load | Handles load without errors |
| `test_03_max_replicas.sh` | 10 replicas stress test | Maximum scale works correctly |
| `test_04_cleanup_verification.sh` | Verify resource cleanup | All Docker resources removed |
| `test_05_long_running.sh` | 5-minute stability test | Server remains stable |
### 4. Dashboard Tests (`dashboard/`)
Dashboard UI functionality with Playwright.
| Test | Description | Expected Result |
|------|-------------|----------------|
| `test_01_dashboard_ui.py` | Full dashboard UI test | All UI elements functional |
**Dashboard Test Details:**
- Starts server with 3 replicas
- Runs demo script to generate activity
- Uses Playwright to:
- Take screenshots of dashboard
- Verify container filter buttons
- Check WebSocket connection
- Validate timeline charts
- Test all dashboard sections
**Screenshots saved to:** `dashboard/screenshots/`
### 5. Edge Case Tests (`edge/`)
Error handling and validation.
| Test | Description | Expected Result |
|------|-------------|----------------|
| `test_01_already_running.sh` | Start when already running | Proper error message |
| `test_02_not_running.sh` | Operations when stopped | Appropriate errors |
| `test_03_scale_single_mode.sh` | Scale single container | Error with guidance |
| `test_04_invalid_port.sh` | Invalid port numbers | Validation errors |
| `test_05_invalid_replicas.sh` | Invalid replica counts | Validation errors |
| `test_06_missing_env_file.sh` | Non-existent env file | File not found error |
| `test_07_port_in_use.sh` | Port already occupied | Port conflict error |
| `test_08_state_corruption.sh` | Corrupted state file | Cleanup recovers |
| `test_09_network_conflict.sh` | Docker network collision | Handles gracefully |
| `test_10_rapid_operations.sh` | Rapid start/stop cycles | No corruption |
## Test Execution Workflow
Each test follows this pattern:
1. **Setup:** Clean state, activate venv
2. **Execute:** Run test commands
3. **Verify:** Check results and assertions
4. **Cleanup:** Stop server, remove resources
## Running Individual Tests
```bash
# Make test executable (if needed)
chmod +x deploy/docker/tests/cli/basic/test_01_start_default.sh
# Run directly
./deploy/docker/tests/cli/basic/test_01_start_default.sh
# Or use the test runner
./run_tests.sh basic 01
```
## Interpreting Results
### Success Output
```
✅ Test passed: [description]
```
### Failure Output
```
❌ Test failed: [error message]
```
### Warning Output
```
⚠️ Warning: [issue description]
```
## Common Issues
### Docker Not Running
```
Error: Docker daemon not running
Solution: Start Docker Desktop or Docker daemon
```
### Port Already In Use
```
Error: Port 11235 is already in use
Solution: Stop existing server or use different port
```
### Virtual Environment Not Found
```
Warning: venv not found
Solution: Create venv and activate it
```
### Playwright Not Installed
```
Error: playwright module not found
Solution: pip install playwright && playwright install chromium
```
## Test Development
### Adding New Tests
1. **Choose category:** basic, advanced, resource, dashboard, or edge
2. **Create test file:** Follow naming pattern `test_XX_description.sh`
3. **Use template:**
```bash
#!/bin/bash
# Test: [Description]
# Expected: [What should happen]
set -e
echo "=== Test: [Name] ==="
echo ""
source venv/bin/activate
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Test logic here
# Cleanup
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: [success message]"
```
4. **Make executable:** `chmod +x test_XX_description.sh`
5. **Test it:** `./test_XX_description.sh`
6. **Add to runner:** Tests are auto-discovered by `run_tests.sh`
## CI/CD Integration
These tests can be integrated into CI/CD pipelines:
```yaml
# Example GitHub Actions
- name: Run CLI Tests
run: |
source venv/bin/activate
cd deploy/docker/tests/cli
./run_tests.sh all
```
## Performance Considerations
- **Basic tests:** ~2-5 minutes total
- **Advanced tests:** ~5-10 minutes total
- **Resource tests:** ~10-15 minutes total (including 5-min stability test)
- **Dashboard test:** ~3-5 minutes
- **Edge case tests:** ~5-8 minutes total
**Full suite:** ~30-45 minutes
## Best Practices
1. **Always cleanup:** Each test should cleanup after itself
2. **Wait for readiness:** Add sleep after starting servers
3. **Check health:** Verify health endpoint before assertions
4. **Graceful failures:** Use `|| true` to continue on expected failures
5. **Clear messages:** Output should clearly indicate what's being tested
## Troubleshooting
### Tests Hanging
- Check if Docker containers are stuck
- Look for port conflicts
- Verify network connectivity
### Intermittent Failures
- Increase sleep durations for slower systems
- Check system resources (memory, CPU)
- Verify Docker has enough resources allocated
### All Tests Failing
- Verify Docker is running: `docker ps`
- Check CLI is installed: `which crwl`
- Activate venv: `source venv/bin/activate`
- Check server manager: `crwl server status`
## Contributing
When adding new tests:
1. Follow existing naming conventions
2. Add comprehensive documentation
3. Test on clean system
4. Update this README
5. Ensure cleanup is robust
## License
Same as Crawl4AI project license.

View File

@@ -1,163 +0,0 @@
# CLI Test Suite - Execution Results
**Date:** 2025-10-20
**Status:** ✅ PASSED
## Summary
| Category | Total | Passed | Failed | Skipped |
|----------|-------|--------|--------|---------|
| Basic Tests | 8 | 8 | 0 | 0 |
| Advanced Tests | 8 | 8 | 0 | 0 |
| Edge Case Tests | 10 | 10 | 0 | 0 |
| Resource Tests | 3 | 3 | 0 | 2 (skipped) |
| Dashboard UI Tests | 0 | 0 | 0 | 1 (not run) |
| **TOTAL** | **29** | **29** | **0** | **3** |
**Success Rate:** 100% (29/29 tests passed)
## Test Results by Category
### ✅ Basic Tests (8/8 Passed)
| Test | Status | Notes |
|------|--------|-------|
| test_01_start_default | ✅ PASS | Server starts with defaults (1 replica, port 11235) |
| test_02_status | ✅ PASS | Status command shows correct information |
| test_03_stop | ✅ PASS | Server stops cleanly, port freed |
| test_04_start_custom_port | ✅ PASS | Server starts on port 8080 |
| test_05_start_replicas | ✅ PASS | Compose mode with 3 replicas |
| test_06_logs | ✅ PASS | Logs retrieved successfully |
| test_07_restart | ✅ PASS | Server restarts preserving config (2 replicas) |
| test_08_cleanup | ✅ PASS | Force cleanup removes all resources |
### ✅ Advanced Tests (8/8 Passed)
| Test | Status | Notes |
|------|--------|-------|
| test_01_scale_up | ✅ PASS | Scaled 3 → 5 replicas successfully |
| test_02_scale_down | ✅ PASS | Scaled 5 → 2 replicas successfully |
| test_03_mode_single | ✅ PASS | Explicit single mode works |
| test_04_mode_compose | ✅ PASS | Compose mode with 3 replicas and Nginx |
| test_05_custom_image | ✅ PASS | Custom image specification works |
| test_06_env_file | ✅ PASS | Environment file loading works |
| test_07_stop_remove_volumes | ✅ PASS | Volumes handled during cleanup |
| test_08_restart_with_scale | ✅ PASS | Restart with scale change (2 → 4 replicas) |
### ✅ Edge Case Tests (10/10 Passed)
| Test | Status | Notes |
|------|--------|-------|
| test_01_already_running | ✅ PASS | Proper error for duplicate start |
| test_02_not_running | ✅ PASS | Appropriate errors when server stopped |
| test_03_scale_single_mode | ✅ PASS | Cannot scale single mode (expected error) |
| test_04_invalid_port | ✅ PASS | Rejected ports: 0, -1, 99999, 65536 |
| test_05_invalid_replicas | ✅ PASS | Rejected replicas: 0, -1, 101 |
| test_06_missing_env_file | ✅ PASS | File not found error |
| test_07_port_in_use | ✅ PASS | Port conflict detected |
| test_08_state_corruption | ✅ PASS | Corrupted state handled gracefully |
| test_09_network_conflict | ✅ PASS | Network collision handled |
| test_10_rapid_operations | ✅ PASS | Rapid start/stop/restart cycles work |
### ✅ Resource Tests (3/5 Completed)
| Test | Status | Notes |
|------|--------|-------|
| test_01_memory_monitoring | ✅ PASS | Baseline: 9.6%, After: 12.1%, Pool: 450 MB |
| test_02_cpu_stress | ✅ PASS | Handled 10 concurrent requests |
| test_03_max_replicas | ⏭️ SKIP | Takes ~2 minutes (10 replicas) |
| test_04_cleanup_verification | ✅ PASS | All resources cleaned up |
| test_05_long_running | ⏭️ SKIP | Takes 5 minutes |
### Dashboard UI Tests (Not Run)
| Test | Status | Notes |
|------|--------|-------|
| test_01_dashboard_ui | ⏭️ SKIP | Requires Playwright, takes ~5 minutes |
## Key Findings
### ✅ Strengths
1. **Robust Error Handling**
- All invalid inputs properly rejected with clear error messages
- State corruption detected and recovered automatically
- Port conflicts identified before container start
2. **Scaling Functionality**
- Live scaling works smoothly (3 → 5 → 2 replicas)
- Mode detection works correctly (single vs compose)
- Restart preserves configuration
3. **Resource Management**
- Cleanup thoroughly removes all Docker resources
- Memory usage reasonable (9.6% → 12.1% with 5 crawls)
- Concurrent requests handled without errors
4. **CLI Usability**
- Clear, color-coded output
- Helpful error messages with hints
- Status command shows comprehensive info
### 📊 Performance Observations
- **Startup Time:** ~5 seconds for single container, ~10-12 seconds for 3 replicas
- **Memory Usage:** Baseline 9.6%, increases to 12.1% after 5 crawls
- **Browser Pool:** ~450 MB memory usage (reasonable)
- **Concurrent Load:** Successfully handled 10 parallel requests
### 🔧 Issues Found
None! All 29 tests passed successfully.
## Test Execution Notes
### Test Environment
- **OS:** macOS (Darwin 24.3.0)
- **Docker:** Running
- **Python:** Virtual environment activated
- **Date:** 2025-10-20
### Skipped Tests Rationale
1. **test_03_max_replicas:** Takes ~2 minutes to start 10 replicas
2. **test_05_long_running:** 5-minute stability test
3. **test_01_dashboard_ui:** Requires Playwright installation, UI screenshots
These tests are fully implemented and can be run manually when time permits.
## Verification Commands
All tests can be re-run with:
```bash
# Individual test
bash deploy/docker/tests/cli/basic/test_01_start_default.sh
# Category
./deploy/docker/tests/cli/run_tests.sh basic
# All tests
./deploy/docker/tests/cli/run_tests.sh all
```
## Conclusion
**The CLI test suite is comprehensive and thoroughly validates all functionality.**
- All core features tested and working
- Error handling is robust
- Edge cases properly covered
- Resource management verified
- No bugs or issues found
The Crawl4AI Docker server CLI is production-ready with excellent test coverage.
---
**Next Steps:**
1. Run skipped tests when time permits (optional)
2. Integrate into CI/CD pipeline
3. Run dashboard UI test for visual verification
4. Document test results in main README
**Recommendation:** ✅ Ready for production use

View File

@@ -1,300 +0,0 @@
# CLI Test Suite - Implementation Summary
## Completed Implementation
Successfully created a comprehensive E2E test suite for the Crawl4AI Docker server CLI.
## Test Suite Overview
### Total Tests: 32
#### 1. Basic Tests (8 tests) ✅
- `test_01_start_default.sh` - Start with default settings
- `test_02_status.sh` - Status command validation
- `test_03_stop.sh` - Clean server shutdown
- `test_04_start_custom_port.sh` - Custom port configuration
- `test_05_start_replicas.sh` - Multi-replica deployment
- `test_06_logs.sh` - Log retrieval
- `test_07_restart.sh` - Server restart
- `test_08_cleanup.sh` - Force cleanup
#### 2. Advanced Tests (8 tests) ✅
- `test_01_scale_up.sh` - Scale from 3 to 5 replicas
- `test_02_scale_down.sh` - Scale from 5 to 2 replicas
- `test_03_mode_single.sh` - Explicit single mode
- `test_04_mode_compose.sh` - Compose mode with Nginx
- `test_05_custom_image.sh` - Custom image specification
- `test_06_env_file.sh` - Environment file loading
- `test_07_stop_remove_volumes.sh` - Volume cleanup
- `test_08_restart_with_scale.sh` - Restart with scale change
#### 3. Resource Tests (5 tests) ✅
- `test_01_memory_monitoring.sh` - Memory usage tracking
- `test_02_cpu_stress.sh` - CPU stress with concurrent requests
- `test_03_max_replicas.sh` - Maximum (10) replicas stress test
- `test_04_cleanup_verification.sh` - Resource cleanup verification
- `test_05_long_running.sh` - 5-minute stability test
#### 4. Dashboard UI Test (1 test) ✅
- `test_01_dashboard_ui.py` - Comprehensive Playwright test
- Automated browser testing
- Screenshot capture (7 screenshots per run)
- UI element validation
- Container filter testing
- WebSocket connection verification
#### 5. Edge Case Tests (10 tests) ✅
- `test_01_already_running.sh` - Duplicate start attempt
- `test_02_not_running.sh` - Operations on stopped server
- `test_03_scale_single_mode.sh` - Invalid scaling operation
- `test_04_invalid_port.sh` - Port validation (0, -1, 99999, 65536)
- `test_05_invalid_replicas.sh` - Replica validation (0, -1, 101)
- `test_06_missing_env_file.sh` - Non-existent env file
- `test_07_port_in_use.sh` - Port conflict detection
- `test_08_state_corruption.sh` - State file corruption recovery
- `test_09_network_conflict.sh` - Docker network collision handling
- `test_10_rapid_operations.sh` - Rapid start/stop cycles
## Test Infrastructure
### Master Test Runner (`run_tests.sh`)
- Run all tests or specific categories
- Color-coded output (green/red/yellow)
- Test counters (passed/failed/skipped)
- Summary statistics
- Individual test execution support
### Documentation
- `README.md` - Comprehensive test documentation
- Test descriptions and expected results
- Usage instructions
- Troubleshooting guide
- Best practices
- CI/CD integration examples
- `TEST_SUMMARY.md` - Implementation summary (this file)
## File Structure
```
deploy/docker/tests/cli/
├── README.md # Main documentation
├── TEST_SUMMARY.md # This summary
├── run_tests.sh # Master test runner
├── basic/ # Basic CLI tests
│ ├── test_01_start_default.sh
│ ├── test_02_status.sh
│ ├── test_03_stop.sh
│ ├── test_04_start_custom_port.sh
│ ├── test_05_start_replicas.sh
│ ├── test_06_logs.sh
│ ├── test_07_restart.sh
│ └── test_08_cleanup.sh
├── advanced/ # Advanced feature tests
│ ├── test_01_scale_up.sh
│ ├── test_02_scale_down.sh
│ ├── test_03_mode_single.sh
│ ├── test_04_mode_compose.sh
│ ├── test_05_custom_image.sh
│ ├── test_06_env_file.sh
│ ├── test_07_stop_remove_volumes.sh
│ └── test_08_restart_with_scale.sh
├── resource/ # Resource and stress tests
│ ├── test_01_memory_monitoring.sh
│ ├── test_02_cpu_stress.sh
│ ├── test_03_max_replicas.sh
│ ├── test_04_cleanup_verification.sh
│ └── test_05_long_running.sh
├── dashboard/ # Dashboard UI tests
│ ├── test_01_dashboard_ui.py
│ ├── run_dashboard_test.sh
│ └── screenshots/ # Auto-generated screenshots
└── edge/ # Edge case tests
├── test_01_already_running.sh
├── test_02_not_running.sh
├── test_03_scale_single_mode.sh
├── test_04_invalid_port.sh
├── test_05_invalid_replicas.sh
├── test_06_missing_env_file.sh
├── test_07_port_in_use.sh
├── test_08_state_corruption.sh
├── test_09_network_conflict.sh
└── test_10_rapid_operations.sh
```
## Usage Examples
### Run All Tests (except dashboard)
```bash
./run_tests.sh
```
### Run Specific Category
```bash
./run_tests.sh basic
./run_tests.sh advanced
./run_tests.sh resource
./run_tests.sh edge
```
### Run Dashboard Tests
```bash
./run_tests.sh dashboard
# or
./dashboard/run_dashboard_test.sh
```
### Run Individual Test
```bash
./run_tests.sh basic 01
./run_tests.sh edge 05
```
### Direct Execution
```bash
./basic/test_01_start_default.sh
./edge/test_01_already_running.sh
```
## Test Verification
The following tests have been verified working:
-`test_01_start_default.sh` - PASSED
-`test_02_status.sh` - PASSED
-`test_03_stop.sh` - PASSED
-`test_03_mode_single.sh` - PASSED
-`test_01_already_running.sh` - PASSED
- ✅ Master test runner - PASSED
## Key Features
### Robustness
- Each test cleans up after itself
- Handles expected failures gracefully
- Waits for server readiness before assertions
- Comprehensive error checking
### Clarity
- Clear test descriptions
- Colored output for easy interpretation
- Detailed error messages
- Progress indicators
### Completeness
- Covers all CLI commands
- Tests success and failure paths
- Validates error messages
- Checks resource cleanup
### Maintainability
- Consistent structure across all tests
- Well-documented code
- Modular test design
- Easy to add new tests
## Test Coverage
### CLI Commands Tested
-`crwl server start` (all options)
-`crwl server stop` (with/without volumes)
-`crwl server status`
-`crwl server scale`
-`crwl server logs`
-`crwl server restart`
-`crwl server cleanup`
### Deployment Modes Tested
- ✅ Single container mode
- ✅ Compose mode (multi-container)
- ✅ Auto mode detection
### Features Tested
- ✅ Custom ports
- ✅ Custom replicas (1-10)
- ✅ Custom images
- ✅ Environment files
- ✅ Live scaling
- ✅ Configuration persistence
- ✅ Resource cleanup
- ✅ Dashboard UI
### Error Handling Tested
- ✅ Invalid inputs (ports, replicas)
- ✅ Missing files
- ✅ Port conflicts
- ✅ State corruption
- ✅ Network conflicts
- ✅ Rapid operations
- ✅ Duplicate operations
## Performance
### Estimated Execution Times
- Basic tests: ~2-5 minutes
- Advanced tests: ~5-10 minutes
- Resource tests: ~10-15 minutes
- Dashboard test: ~3-5 minutes
- Edge case tests: ~5-8 minutes
**Total: ~30-45 minutes for full suite**
## Next Steps
### Recommended Actions
1. ✅ Run full test suite to verify all tests
2. ✅ Test dashboard UI test with Playwright
3. ✅ Verify long-running stability test
4. ✅ Integrate into CI/CD pipeline
5. ✅ Add to project documentation
### Future Enhancements
- Add performance benchmarking
- Add load testing scenarios
- Add network failure simulation
- Add disk space tests
- Add security tests
- Add multi-host tests (Swarm mode)
## Notes
### Dependencies
- Docker running
- Virtual environment activated
- `jq` for JSON parsing (installed by default on most systems)
- `bc` for calculations (installed by default on most systems)
- Playwright for dashboard tests (optional)
### Test Philosophy
- **Small:** Each test focuses on one specific aspect
- **Smart:** Tests verify both success and failure paths
- **Strong:** Robust cleanup and error handling
- **Self-contained:** Each test is independent
### Known Limitations
- Dashboard test requires Playwright installation
- Long-running test takes 5 minutes
- Max replicas test requires significant system resources
- Some tests may need adjustment for slower systems
## Success Criteria
✅ All 32 tests created
✅ Test runner implemented
✅ Documentation complete
✅ Tests verified working
✅ File structure organized
✅ Error handling comprehensive
✅ Cleanup mechanisms robust
## Conclusion
The CLI test suite is complete and ready for use. It provides comprehensive coverage of all CLI functionality, validates error handling, and ensures robustness across various scenarios.
**Status:** ✅ COMPLETE
**Date:** 2025-10-20
**Tests:** 32 (8 basic + 8 advanced + 5 resource + 1 dashboard + 10 edge)

View File

@@ -1,56 +0,0 @@
#!/bin/bash
# Test: Scale server up from 3 to 5 replicas
# Expected: Server scales without downtime
set -e
echo "=== Test: Scale Up (3 → 5 replicas) ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with 3 replicas
echo "Starting server with 3 replicas..."
crwl server start --replicas 3 >/dev/null 2>&1
sleep 10
# Verify 3 replicas
STATUS=$(crwl server status | grep "Replicas" || echo "")
echo "Initial status: $STATUS"
# Scale up to 5
echo ""
echo "Scaling up to 5 replicas..."
crwl server scale 5
sleep 10
# Verify 5 replicas
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "5"; then
echo "❌ Status does not show 5 replicas"
crwl server stop
exit 1
fi
# Verify health during scaling
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after scaling"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Successfully scaled from 3 to 5 replicas"

View File

@@ -1,56 +0,0 @@
#!/bin/bash
# Test: Scale server down from 5 to 2 replicas
# Expected: Server scales down gracefully
set -e
echo "=== Test: Scale Down (5 → 2 replicas) ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with 5 replicas
echo "Starting server with 5 replicas..."
crwl server start --replicas 5 >/dev/null 2>&1
sleep 12
# Verify 5 replicas
STATUS=$(crwl server status | grep "Replicas" || echo "")
echo "Initial status: $STATUS"
# Scale down to 2
echo ""
echo "Scaling down to 2 replicas..."
crwl server scale 2
sleep 8
# Verify 2 replicas
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "2"; then
echo "❌ Status does not show 2 replicas"
crwl server stop
exit 1
fi
# Verify health after scaling down
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after scaling down"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Successfully scaled down from 5 to 2 replicas"

View File

@@ -1,52 +0,0 @@
#!/bin/bash
# Test: Start server explicitly in single mode
# Expected: Server starts in single mode
set -e
echo "=== Test: Explicit Single Mode ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start in single mode explicitly
echo "Starting server in single mode..."
crwl server start --mode single
sleep 5
# Check mode
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "single"; then
echo "❌ Mode is not 'single'"
crwl server stop
exit 1
fi
if ! echo "$STATUS" | grep -q "1"; then
echo "❌ Should have 1 replica in single mode"
crwl server stop
exit 1
fi
# Verify health
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server started in single mode"

View File

@@ -1,52 +0,0 @@
#!/bin/bash
# Test: Start server in compose mode with replicas
# Expected: Server starts in compose mode with Nginx
set -e
echo "=== Test: Compose Mode with 3 Replicas ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start in compose mode
echo "Starting server in compose mode with 3 replicas..."
crwl server start --mode compose --replicas 3
sleep 12
# Check mode
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "3"; then
echo "❌ Status does not show 3 replicas"
crwl server stop
exit 1
fi
# Verify Nginx is running (load balancer)
NGINX_RUNNING=$(docker ps --filter "name=nginx" --format "{{.Names}}" || echo "")
if [[ -z "$NGINX_RUNNING" ]]; then
echo "⚠️ Warning: Nginx load balancer not detected (may be using swarm or single mode)"
fi
# Verify health through load balancer
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server started in compose mode"

View File

@@ -1,47 +0,0 @@
#!/bin/bash
# Test: Start server with custom image tag
# Expected: Server uses specified image
set -e
echo "=== Test: Custom Image Specification ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Use latest tag explicitly (or specify a different tag if available)
IMAGE="unclecode/crawl4ai:latest"
echo "Starting server with image: $IMAGE..."
crwl server start --image "$IMAGE"
sleep 5
# Check status shows correct image
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "crawl4ai"; then
echo "❌ Status does not show correct image"
crwl server stop
exit 1
fi
# Verify health
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server started with custom image"

View File

@@ -1,47 +0,0 @@
#!/bin/bash
# Test: Start server with environment file
# Expected: Server loads environment variables
set -e
echo "=== Test: Start with Environment File ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Create a test env file
TEST_ENV_FILE="/tmp/test_crawl4ai.env"
cat > "$TEST_ENV_FILE" <<EOF
TEST_VAR=test_value
OPENAI_API_KEY=sk-test-key
EOF
echo "Created test env file at $TEST_ENV_FILE"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with env file
echo "Starting server with env file..."
crwl server start --env-file "$TEST_ENV_FILE"
sleep 5
# Verify server started
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed"
rm -f "$TEST_ENV_FILE"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
rm -f "$TEST_ENV_FILE"
echo ""
echo "✅ Test passed: Server started with environment file"

View File

@@ -1,49 +0,0 @@
#!/bin/bash
# Test: Stop server with volume removal
# Expected: Volumes are removed along with containers
set -e
echo "=== Test: Stop with Remove Volumes ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server (which may create volumes)
echo "Starting server..."
crwl server start --replicas 2 >/dev/null 2>&1
sleep 8
# Make some requests to populate data
echo "Making requests to populate data..."
curl -s -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d '{"urls": ["https://httpbin.org/html"], "crawler_config": {}}' > /dev/null || true
sleep 2
# Stop with volume removal (needs confirmation, so we'll use cleanup instead)
echo "Stopping server with volume removal..."
# Note: --remove-volumes requires confirmation, so we use cleanup --force
crwl server cleanup --force >/dev/null 2>&1
sleep 3
# Verify volumes are removed
echo "Checking for remaining volumes..."
VOLUMES=$(docker volume ls --filter "name=crawl4ai" --format "{{.Name}}" || echo "")
if [[ -n "$VOLUMES" ]]; then
echo "⚠️ Warning: Some volumes still exist: $VOLUMES"
echo " (This may be expected if using system-wide volumes)"
fi
# Verify server is stopped
STATUS=$(crwl server status | grep "No server" || echo "RUNNING")
if [[ "$STATUS" == "RUNNING" ]]; then
echo "❌ Server still running after stop"
exit 1
fi
echo ""
echo "✅ Test passed: Server stopped and volumes handled"

View File

@@ -1,56 +0,0 @@
#!/bin/bash
# Test: Restart server with different replica count
# Expected: Server restarts with new replica count
set -e
echo "=== Test: Restart with Scale Change ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with 2 replicas
echo "Starting server with 2 replicas..."
crwl server start --replicas 2 >/dev/null 2>&1
sleep 8
# Verify 2 replicas
STATUS=$(crwl server status | grep "Replicas" || echo "")
echo "Initial: $STATUS"
# Restart with 4 replicas
echo ""
echo "Restarting with 4 replicas..."
crwl server restart --replicas 4
sleep 10
# Verify 4 replicas
STATUS=$(crwl server status)
echo "$STATUS"
if ! echo "$STATUS" | grep -q "4"; then
echo "❌ Status does not show 4 replicas after restart"
crwl server stop
exit 1
fi
# Verify health
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after restart"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server restarted with new replica count"

View File

@@ -1,52 +0,0 @@
#!/bin/bash
# Test: Start server with default settings
# Expected: Server starts with 1 replica on port 11235
set -e
echo "=== Test: Start Server with Defaults ==="
echo "Expected: 1 replica, port 11235, auto mode"
echo ""
# Activate virtual environment
# Navigate to project root and activate venv
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup any existing server
echo "Cleaning up any existing server..."
crwl server stop 2>/dev/null || true
sleep 2
# Start server with defaults
echo "Starting server with default settings..."
crwl server start
# Wait for server to be ready
echo "Waiting for server to be healthy..."
sleep 5
# Verify server is running
echo "Checking server status..."
STATUS=$(crwl server status | grep "Running" || echo "NOT_RUNNING")
if [[ "$STATUS" == "NOT_RUNNING" ]]; then
echo "❌ Server failed to start"
crwl server stop
exit 1
fi
# Check health endpoint
echo "Checking health endpoint..."
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed: $HEALTH"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop
echo ""
echo "✅ Test passed: Server started with defaults and responded to health check"

View File

@@ -1,42 +0,0 @@
#!/bin/bash
# Test: Check server status command
# Expected: Shows running status with correct details
set -e
echo "=== Test: Server Status Command ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server first
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Check status
echo "Checking server status..."
STATUS_OUTPUT=$(crwl server status)
echo "$STATUS_OUTPUT"
echo ""
# Verify output contains expected fields
if ! echo "$STATUS_OUTPUT" | grep -q "Running"; then
echo "❌ Status does not show 'Running'"
crwl server stop
exit 1
fi
if ! echo "$STATUS_OUTPUT" | grep -q "11235"; then
echo "❌ Status does not show correct port"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Status command shows correct information"

View File

@@ -1,45 +0,0 @@
#!/bin/bash
# Test: Stop server command
# Expected: Server stops cleanly and port becomes available
set -e
echo "=== Test: Stop Server Command ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server first
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Verify running
echo "Verifying server is running..."
if ! curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Server is not running before stop"
exit 1
fi
# Stop server
echo "Stopping server..."
crwl server stop
# Verify stopped
echo "Verifying server is stopped..."
sleep 3
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Server is still responding after stop"
exit 1
fi
# Check status shows not running
STATUS=$(crwl server status | grep "No server" || echo "RUNNING")
if [[ "$STATUS" == "RUNNING" ]]; then
echo "❌ Status still shows server as running"
exit 1
fi
echo ""
echo "✅ Test passed: Server stopped cleanly"

View File

@@ -1,46 +0,0 @@
#!/bin/bash
# Test: Start server with custom port
# Expected: Server starts on port 8080 instead of default 11235
set -e
echo "=== Test: Start Server with Custom Port ==="
echo "Expected: Server on port 8080"
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start on custom port
echo "Starting server on port 8080..."
crwl server start --port 8080
sleep 5
# Check health on custom port
echo "Checking health on port 8080..."
HEALTH=$(curl -s http://localhost:8080/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed on port 8080: $HEALTH"
crwl server stop
exit 1
fi
# Verify default port is NOT responding
echo "Verifying port 11235 is not in use..."
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Server is also running on default port 11235"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop
echo ""
echo "✅ Test passed: Server started on custom port 8080"

View File

@@ -1,54 +0,0 @@
#!/bin/bash
# Test: Start server with multiple replicas
# Expected: Server starts with 3 replicas in compose mode
set -e
echo "=== Test: Start Server with 3 Replicas ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with 3 replicas
echo "Starting server with 3 replicas..."
crwl server start --replicas 3
sleep 10
# Check status shows 3 replicas
echo "Checking status..."
STATUS_OUTPUT=$(crwl server status)
echo "$STATUS_OUTPUT"
if ! echo "$STATUS_OUTPUT" | grep -q "3"; then
echo "❌ Status does not show 3 replicas"
crwl server stop
exit 1
fi
# Check health endpoint
echo "Checking health endpoint..."
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed"
crwl server stop
exit 1
fi
# Check container discovery (should show 3 containers eventually)
echo "Checking container discovery..."
sleep 5 # Wait for heartbeats
CONTAINERS=$(curl -s http://localhost:11235/monitor/containers | jq -r '.count' 2>/dev/null || echo "0")
echo "Container count: $CONTAINERS"
# Cleanup
echo "Cleaning up..."
crwl server stop
echo ""
echo "✅ Test passed: Server started with 3 replicas"

View File

@@ -1,47 +0,0 @@
#!/bin/bash
# Test: View server logs
# Expected: Logs are displayed without errors
set -e
echo "=== Test: Server Logs Command ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Make a request to generate some logs
echo "Making request to generate logs..."
curl -s http://localhost:11235/health > /dev/null
# Check logs (tail)
echo "Fetching logs (last 50 lines)..."
LOGS=$(crwl server logs --tail 50 2>&1 || echo "ERROR")
if [[ "$LOGS" == "ERROR" ]]; then
echo "❌ Failed to retrieve logs"
crwl server stop
exit 1
fi
echo "Log sample (first 10 lines):"
echo "$LOGS" | head -n 10
echo ""
# Verify logs contain something (not empty)
if [[ -z "$LOGS" ]]; then
echo "❌ Logs are empty"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Logs retrieved successfully"

View File

@@ -1,55 +0,0 @@
#!/bin/bash
# Test: Restart server command
# Expected: Server restarts with same configuration
set -e
echo "=== Test: Restart Server Command ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server with specific config
echo "Starting server with 2 replicas..."
crwl server start --replicas 2 >/dev/null 2>&1
sleep 8
# Get initial container ID
echo "Getting initial state..."
INITIAL_STATUS=$(crwl server status)
echo "$INITIAL_STATUS"
# Restart
echo ""
echo "Restarting server..."
crwl server restart
sleep 8
# Check status after restart
echo "Checking status after restart..."
RESTART_STATUS=$(crwl server status)
echo "$RESTART_STATUS"
# Verify still has 2 replicas
if ! echo "$RESTART_STATUS" | grep -q "2"; then
echo "❌ Replica count not preserved after restart"
crwl server stop
exit 1
fi
# Verify health
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after restart"
crwl server stop
exit 1
fi
# Cleanup
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server restarted with preserved configuration"

View File

@@ -1,46 +0,0 @@
#!/bin/bash
# Test: Force cleanup command
# Expected: All resources removed even if state is corrupted
set -e
echo "=== Test: Force Cleanup Command ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Run cleanup (will prompt, so use force flag)
echo "Running force cleanup..."
crwl server cleanup --force
sleep 3
# Verify no containers running
echo "Verifying cleanup..."
CONTAINERS=$(docker ps --filter "name=crawl4ai" --format "{{.Names}}" || echo "")
if [[ -n "$CONTAINERS" ]]; then
echo "❌ Crawl4AI containers still running: $CONTAINERS"
exit 1
fi
# Verify port is free
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Server still responding after cleanup"
exit 1
fi
# Verify status shows not running
STATUS=$(crwl server status | grep "No server" || echo "RUNNING")
if [[ "$STATUS" == "RUNNING" ]]; then
echo "❌ Status still shows server running after cleanup"
exit 1
fi
echo ""
echo "✅ Test passed: Force cleanup removed all resources"

View File

@@ -1,27 +0,0 @@
#!/bin/bash
# Wrapper script to run dashboard UI test with proper environment
set -e
echo "=== Dashboard UI Test ==="
echo ""
# Activate virtual environment
source venv/bin/activate
# Make sure playwright is installed
echo "Checking Playwright installation..."
python -c "import playwright" 2>/dev/null || {
echo "Installing Playwright..."
pip install playwright
playwright install chromium
}
# Run the test
echo ""
echo "Running dashboard UI test..."
python deploy/docker/tests/cli/dashboard/test_01_dashboard_ui.py
echo ""
echo "✅ Dashboard test complete"
echo "Check deploy/docker/tests/cli/dashboard/screenshots/ for results"

View File

@@ -1,225 +0,0 @@
#!/usr/bin/env python3
"""
Dashboard UI Test with Playwright
Tests the monitoring dashboard UI functionality
"""
import asyncio
import subprocess
import time
import os
from pathlib import Path
from playwright.async_api import async_playwright
BASE_URL = "http://localhost:11235"
SCREENSHOT_DIR = Path(__file__).parent / "screenshots"
async def start_server():
"""Start server with 3 replicas"""
print("Starting server with 3 replicas...")
subprocess.run(["crwl", "server", "stop"],
stdout=subprocess.DEVNULL,
stderr=subprocess.DEVNULL)
time.sleep(2)
result = subprocess.run(
["crwl", "server", "start", "--replicas", "3"],
capture_output=True,
text=True
)
if result.returncode != 0:
raise Exception(f"Failed to start server: {result.stderr}")
print("Waiting for server to be ready...")
time.sleep(12)
async def run_demo_script():
"""Run the demo script in background to generate activity"""
print("Starting demo script to generate dashboard activity...")
demo_path = Path(__file__).parent.parent.parent / "monitor" / "demo_monitor_dashboard.py"
process = subprocess.Popen(
["python", str(demo_path)],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE
)
# Let it run for a bit to generate some data
print("Waiting for demo to generate data...")
time.sleep(10)
return process
async def test_dashboard_ui():
"""Test dashboard UI with Playwright"""
# Create screenshot directory
SCREENSHOT_DIR.mkdir(exist_ok=True)
print(f"Screenshots will be saved to: {SCREENSHOT_DIR}")
async with async_playwright() as p:
# Launch browser
print("\nLaunching browser...")
browser = await p.chromium.launch(headless=True)
context = await browser.new_context(viewport={'width': 1920, 'height': 1080})
page = await context.new_page()
try:
# Navigate to dashboard
print(f"Navigating to {BASE_URL}/dashboard")
await page.goto(f"{BASE_URL}/dashboard", wait_until="networkidle")
await asyncio.sleep(3)
# Take full dashboard screenshot
print("Taking full dashboard screenshot...")
await page.screenshot(path=SCREENSHOT_DIR / "01_full_dashboard.png", full_page=True)
print(f" ✅ Saved: 01_full_dashboard.png")
# Verify page title
title = await page.title()
print(f"\nPage title: {title}")
if "Monitor" not in title and "Dashboard" not in title:
print(" ⚠️ Warning: Title doesn't contain 'Monitor' or 'Dashboard'")
# Check for infrastructure card (container filters)
print("\nChecking Infrastructure card...")
infrastructure = await page.query_selector('.card h3:has-text("Infrastructure")')
if infrastructure:
print(" ✅ Infrastructure card found")
await page.screenshot(path=SCREENSHOT_DIR / "02_infrastructure_card.png")
print(f" ✅ Saved: 02_infrastructure_card.png")
else:
print(" ❌ Infrastructure card not found")
# Check for container filter buttons (All, C-1, C-2, C-3)
print("\nChecking container filter buttons...")
all_button = await page.query_selector('.filter-btn[data-container="all"]')
if all_button:
print("'All' filter button found")
# Take screenshot of filter area
await all_button.screenshot(path=SCREENSHOT_DIR / "03_filter_buttons.png")
print(f" ✅ Saved: 03_filter_buttons.png")
# Test clicking filter button
await all_button.click()
await asyncio.sleep(1)
print(" ✅ Clicked 'All' filter button")
else:
print(" ⚠️ 'All' filter button not found (may appear after containers register)")
# Check for WebSocket connection indicator
print("\nChecking WebSocket connection...")
ws_indicator = await page.query_selector('.ws-status, .connection-status, [class*="websocket"]')
if ws_indicator:
print(" ✅ WebSocket indicator found")
else:
print(" ⚠️ WebSocket indicator not found in DOM")
# Check for main dashboard sections
print("\nChecking dashboard sections...")
sections = [
("Active Requests", ".active-requests, [class*='active']"),
("Completed Requests", ".completed-requests, [class*='completed']"),
("Browsers", ".browsers, [class*='browser']"),
("Timeline", ".timeline, [class*='timeline']"),
]
for section_name, selector in sections:
element = await page.query_selector(selector)
if element:
print(f"{section_name} section found")
else:
print(f" ⚠️ {section_name} section not found with selector: {selector}")
# Scroll to different sections and take screenshots
print("\nTaking section screenshots...")
# Requests section
requests = await page.query_selector('.card h3:has-text("Requests")')
if requests:
await requests.scroll_into_view_if_needed()
await asyncio.sleep(1)
await page.screenshot(path=SCREENSHOT_DIR / "04_requests_section.png")
print(f" ✅ Saved: 04_requests_section.png")
# Browsers section
browsers = await page.query_selector('.card h3:has-text("Browsers")')
if browsers:
await browsers.scroll_into_view_if_needed()
await asyncio.sleep(1)
await page.screenshot(path=SCREENSHOT_DIR / "05_browsers_section.png")
print(f" ✅ Saved: 05_browsers_section.png")
# Timeline section
timeline = await page.query_selector('.card h3:has-text("Timeline")')
if timeline:
await timeline.scroll_into_view_if_needed()
await asyncio.sleep(1)
await page.screenshot(path=SCREENSHOT_DIR / "06_timeline_section.png")
print(f" ✅ Saved: 06_timeline_section.png")
# Check for tabs (if they exist)
print("\nChecking for tabs...")
tabs = await page.query_selector_all('.tab, [role="tab"]')
if tabs:
print(f" ✅ Found {len(tabs)} tabs")
for i, tab in enumerate(tabs[:5]): # Check first 5 tabs
tab_text = await tab.inner_text()
print(f" - Tab {i+1}: {tab_text}")
else:
print(" No tab elements found")
# Wait for any animations to complete
await asyncio.sleep(2)
# Take final screenshot
print("\nTaking final screenshot...")
await page.screenshot(path=SCREENSHOT_DIR / "07_final_state.png", full_page=True)
print(f" ✅ Saved: 07_final_state.png")
print("\n" + "="*60)
print("Dashboard UI Test Complete!")
print(f"Screenshots saved to: {SCREENSHOT_DIR}")
print("="*60)
finally:
await browser.close()
async def cleanup():
"""Stop server and cleanup"""
print("\nCleaning up...")
subprocess.run(["crwl", "server", "stop"],
stdout=subprocess.DEVNULL,
stderr=subprocess.DEVNULL)
print("✅ Cleanup complete")
async def main():
"""Main test execution"""
demo_process = None
try:
# Start server
await start_server()
# Run demo script to generate activity
demo_process = await run_demo_script()
# Run dashboard UI test
await test_dashboard_ui()
print("\n✅ All dashboard UI tests passed!")
except Exception as e:
print(f"\n❌ Test failed: {e}")
raise
finally:
# Stop demo script
if demo_process:
demo_process.terminate()
demo_process.wait(timeout=5)
# Cleanup server
await cleanup()
if __name__ == "__main__":
asyncio.run(main())

View File

@@ -1,48 +0,0 @@
#!/bin/bash
# Test: Try starting server when already running
# Expected: Error message indicating server is already running
set -e
echo "=== Test: Start When Already Running ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start server
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Try to start again
echo ""
echo "Attempting to start server again (should fail)..."
OUTPUT=$(crwl server start 2>&1 || true)
echo "$OUTPUT"
# Verify error message
if echo "$OUTPUT" | grep -iq "already running"; then
echo ""
echo "✅ Test passed: Proper error for already running server"
else
echo ""
echo "❌ Test failed: Expected 'already running' error message"
crwl server stop
exit 1
fi
# Verify original server still running
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Original server is not running"
crwl server stop
exit 1
fi
# Cleanup
crwl server stop >/dev/null 2>&1

View File

@@ -1,50 +0,0 @@
#!/bin/bash
# Test: Operations when server is not running
# Expected: Appropriate error messages
set -e
echo "=== Test: Operations When Not Running ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Make sure nothing is running
crwl server stop 2>/dev/null || true
sleep 2
# Try status when not running
echo "Checking status when not running..."
OUTPUT=$(crwl server status 2>&1 || true)
echo "$OUTPUT"
echo ""
if ! echo "$OUTPUT" | grep -iq "no server"; then
echo "❌ Status should indicate no server running"
exit 1
fi
# Try stop when not running
echo "Trying to stop when not running..."
OUTPUT=$(crwl server stop 2>&1 || true)
echo "$OUTPUT"
echo ""
if ! echo "$OUTPUT" | grep -iq "no server\|not running"; then
echo "❌ Stop should indicate no server running"
exit 1
fi
# Try scale when not running
echo "Trying to scale when not running..."
OUTPUT=$(crwl server scale 3 2>&1 || true)
echo "$OUTPUT"
echo ""
if ! echo "$OUTPUT" | grep -iq "no server\|not running"; then
echo "❌ Scale should indicate no server running"
exit 1
fi
echo "✅ Test passed: Appropriate errors for operations when not running"

View File

@@ -1,47 +0,0 @@
#!/bin/bash
# Test: Try to scale single container mode
# Expected: Error indicating single mode cannot be scaled
set -e
echo "=== Test: Scale Single Container Mode ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start in single mode
echo "Starting in single mode..."
crwl server start --mode single >/dev/null 2>&1
sleep 5
# Try to scale
echo ""
echo "Attempting to scale single mode (should fail)..."
OUTPUT=$(crwl server scale 3 2>&1 || true)
echo "$OUTPUT"
echo ""
# Verify error message
if echo "$OUTPUT" | grep -iq "single"; then
echo "✅ Test passed: Proper error for scaling single mode"
else
echo "❌ Test failed: Expected error about single mode"
crwl server stop
exit 1
fi
# Verify server still running
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Server is not running after failed scale"
crwl server stop
exit 1
fi
# Cleanup
crwl server stop >/dev/null 2>&1

View File

@@ -1,36 +0,0 @@
#!/bin/bash
# Test: Invalid port numbers
# Expected: Validation errors for invalid ports
set -e
echo "=== Test: Invalid Port Numbers ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Test invalid ports
INVALID_PORTS=(0 -1 99999 65536)
for PORT in "${INVALID_PORTS[@]}"; do
echo "Testing invalid port: $PORT"
OUTPUT=$(crwl server start --port $PORT 2>&1 || true)
if echo "$OUTPUT" | grep -iq "error\|invalid\|usage"; then
echo " ✅ Rejected port $PORT"
else
echo " ⚠️ Port $PORT may have been accepted (output: $OUTPUT)"
fi
# Make sure no server started
crwl server stop 2>/dev/null || true
sleep 1
echo ""
done
echo "✅ Test passed: Invalid ports handled appropriately"

View File

@@ -1,57 +0,0 @@
#!/bin/bash
# Test: Invalid replica counts
# Expected: Validation errors for invalid replicas
set -e
echo "=== Test: Invalid Replica Counts ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Test invalid replica counts
INVALID_REPLICAS=(0 -1 101)
for REPLICAS in "${INVALID_REPLICAS[@]}"; do
echo "Testing invalid replica count: $REPLICAS"
OUTPUT=$(crwl server start --replicas $REPLICAS 2>&1 || true)
if echo "$OUTPUT" | grep -iq "error\|invalid\|usage"; then
echo " ✅ Rejected replica count $REPLICAS"
else
echo " ⚠️ Replica count $REPLICAS may have been accepted"
fi
# Make sure no server started
crwl server stop 2>/dev/null || true
sleep 1
echo ""
done
# Test scaling to invalid counts
echo "Testing scale to invalid counts..."
crwl server start --replicas 2 >/dev/null 2>&1
sleep 5
INVALID_SCALE=(0 -1)
for SCALE in "${INVALID_SCALE[@]}"; do
echo "Testing scale to: $SCALE"
OUTPUT=$(crwl server scale $SCALE 2>&1 || true)
if echo "$OUTPUT" | grep -iq "error\|invalid\|must be at least 1"; then
echo " ✅ Rejected scale to $SCALE"
else
echo " ⚠️ Scale to $SCALE may have been accepted"
fi
echo ""
done
# Cleanup
crwl server stop >/dev/null 2>&1
echo "✅ Test passed: Invalid replica counts handled appropriately"

View File

@@ -1,40 +0,0 @@
#!/bin/bash
# Test: Non-existent environment file
# Expected: Error indicating file not found
set -e
echo "=== Test: Missing Environment File ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Try with non-existent file
FAKE_FILE="/tmp/nonexistent_$(date +%s).env"
echo "Attempting to start with non-existent env file: $FAKE_FILE"
OUTPUT=$(crwl server start --env-file "$FAKE_FILE" 2>&1 || true)
echo "$OUTPUT"
echo ""
# Verify error
if echo "$OUTPUT" | grep -iq "error\|does not exist\|not found\|no such file"; then
echo "✅ Test passed: Proper error for missing env file"
else
echo "❌ Test failed: Expected error about missing file"
crwl server stop
exit 1
fi
# Make sure no server started
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Server should not have started"
crwl server stop
exit 1
fi
echo "✅ Server correctly refused to start with missing env file"

View File

@@ -1,50 +0,0 @@
#!/bin/bash
# Test: Port already in use
# Expected: Error indicating port is occupied
set -e
echo "=== Test: Port Already In Use ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start a simple HTTP server on port 11235 to occupy it
echo "Starting dummy server on port 11235..."
python -m http.server 11235 >/dev/null 2>&1 &
DUMMY_PID=$!
sleep 2
# Try to start crawl4ai on same port
echo "Attempting to start Crawl4AI on occupied port..."
OUTPUT=$(crwl server start 2>&1 || true)
echo "$OUTPUT"
echo ""
# Kill dummy server
kill $DUMMY_PID 2>/dev/null || true
sleep 1
# Verify error message
if echo "$OUTPUT" | grep -iq "port.*in use\|already in use\|address already in use"; then
echo "✅ Test passed: Proper error for port in use"
else
echo "⚠️ Expected 'port in use' error (output may vary)"
fi
# Make sure Crawl4AI didn't start
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "unknown")
if [[ "$HEALTH" == "ok" ]]; then
echo "❌ Crawl4AI started despite port being occupied"
crwl server stop
exit 1
fi
fi
echo "✅ Crawl4AI correctly refused to start on occupied port"

View File

@@ -1,79 +0,0 @@
#!/bin/bash
# Test: Corrupted state file
# Expected: Cleanup recovers from corrupted state
set -e
echo "=== Test: State File Corruption ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start server to create state
echo "Starting server to create state..."
crwl server start >/dev/null 2>&1
sleep 5
# Get state file path
STATE_FILE="$HOME/.crawl4ai/server/state.json"
echo "State file: $STATE_FILE"
# Verify state file exists
if [[ ! -f "$STATE_FILE" ]]; then
echo "❌ State file not created"
crwl server stop
exit 1
fi
echo "Original state:"
cat "$STATE_FILE" | jq '.' || cat "$STATE_FILE"
echo ""
# Stop server
crwl server stop >/dev/null 2>&1
sleep 2
# Corrupt state file
echo "Corrupting state file..."
echo "{ invalid json }" > "$STATE_FILE"
cat "$STATE_FILE"
echo ""
# Try to start server (should handle corrupted state)
echo "Attempting to start with corrupted state..."
OUTPUT=$(crwl server start 2>&1 || true)
echo "$OUTPUT"
echo ""
# Check if server started or gave clear error
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "✅ Server started despite corrupted state"
crwl server stop
elif echo "$OUTPUT" | grep -iq "already running"; then
# State thinks server is running, use cleanup
echo "State thinks server is running, using cleanup..."
crwl server cleanup --force >/dev/null 2>&1
sleep 2
# Try starting again
crwl server start >/dev/null 2>&1
sleep 5
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "✅ Cleanup recovered from corrupted state"
crwl server stop
else
echo "❌ Failed to recover from corrupted state"
exit 1
fi
else
echo "✅ Handled corrupted state appropriately"
fi
echo ""
echo "✅ Test passed: System handles state corruption"

View File

@@ -1,47 +0,0 @@
#!/bin/bash
# Test: Docker network name collision
# Expected: Handles existing network gracefully
set -e
echo "=== Test: Network Name Conflict ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Create a network with similar name
NETWORK_NAME="crawl4ai_test_net"
echo "Creating test network: $NETWORK_NAME..."
docker network create "$NETWORK_NAME" 2>/dev/null || echo "Network may already exist"
# Start server (should either use existing network or create its own)
echo ""
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Verify server started successfully
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Server failed to start"
docker network rm "$NETWORK_NAME" 2>/dev/null || true
crwl server stop
exit 1
fi
echo "✅ Server started successfully despite network conflict"
# Cleanup
crwl server stop >/dev/null 2>&1
sleep 2
# Remove test network
docker network rm "$NETWORK_NAME" 2>/dev/null || echo "Network already removed"
echo ""
echo "✅ Test passed: Handled network conflict gracefully"

View File

@@ -1,72 +0,0 @@
#!/bin/bash
# Test: Rapid start/stop/restart operations
# Expected: System handles rapid operations without corruption
set -e
echo "=== Test: Rapid Operations ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Test 1: Rapid start/stop
echo "Test 1: Rapid start/stop cycles..."
for i in {1..3}; do
echo " Cycle $i/3..."
crwl server start >/dev/null 2>&1
sleep 3
crwl server stop >/dev/null 2>&1
sleep 2
done
echo " ✅ Completed rapid start/stop cycles"
# Test 2: Restart immediately after start
echo ""
echo "Test 2: Restart immediately after start..."
crwl server start >/dev/null 2>&1
sleep 3
crwl server restart >/dev/null 2>&1
sleep 5
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo " ❌ Health check failed after rapid restart"
crwl server stop
exit 1
fi
echo " ✅ Rapid restart successful"
# Test 3: Multiple status checks
echo ""
echo "Test 3: Multiple rapid status checks..."
for i in {1..5}; do
crwl server status >/dev/null 2>&1 || echo " ⚠️ Status check $i failed"
done
echo " ✅ Multiple status checks completed"
# Test 4: Stop and immediate start
echo ""
echo "Test 4: Stop and immediate start..."
crwl server stop >/dev/null 2>&1
sleep 2
crwl server start >/dev/null 2>&1
sleep 5
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo " ❌ Health check failed after stop/start"
crwl server stop
exit 1
fi
echo " ✅ Stop/immediate start successful"
# Cleanup
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: System handles rapid operations correctly"

View File

@@ -1,119 +0,0 @@
E2E CLI Test Suite Plan │ │
│ │ │ │
│ │ Test Structure │ │
│ │ │ │
│ │ Create deploy/docker/tests/cli/ folder with individual test scripts organized by category. │ │
│ │ │ │
│ │ Test Categories │ │
│ │ │ │
│ │ 1. Basic Tests (deploy/docker/tests/cli/basic/) │ │
│ │ │ │
│ │ - test_01_start_default.sh - Start server with defaults (1 replica, port 11235) │ │
│ │ - test_02_status.sh - Check server status │ │
│ │ - test_03_stop.sh - Stop server cleanly │ │
│ │ - test_04_start_custom_port.sh - Start with custom port (8080) │ │
│ │ - test_05_start_replicas.sh - Start with 3 replicas │ │
│ │ - test_06_logs.sh - View logs (tail and follow) │ │
│ │ - test_07_restart.sh - Restart server preserving config │ │
│ │ - test_08_cleanup.sh - Force cleanup all resources │ │
│ │ │ │
│ │ 2. Advanced Tests (deploy/docker/tests/cli/advanced/) │ │
│ │ │ │
│ │ - test_01_scale_up.sh - Scale from 3 to 5 replicas │ │
│ │ - test_02_scale_down.sh - Scale from 5 to 2 replicas │ │
│ │ - test_03_mode_single.sh - Start in single mode explicitly │ │
│ │ - test_04_mode_compose.sh - Start in compose mode with 3 replicas │ │
│ │ - test_05_custom_image.sh - Start with custom image tag │ │
│ │ - test_06_env_file.sh - Start with custom env file │ │
│ │ - test_07_stop_remove_volumes.sh - Stop and remove volumes │ │
│ │ - test_08_restart_with_scale.sh - Restart and change replica count │ │
│ │ │ │
│ │ 3. Resource Tests (deploy/docker/tests/cli/resource/) │ │
│ │ │ │
│ │ - test_01_memory_monitoring.sh - Monitor memory during crawls │ │
│ │ - test_02_cpu_stress.sh - CPU usage under concurrent load │ │
│ │ - test_03_max_replicas.sh - Start with 10 replicas and stress test │ │
│ │ - test_04_cleanup_verification.sh - Verify all resources cleaned up │ │
│ │ - test_05_long_running.sh - Stability test (30 min runtime) │ │
│ │ │ │
│ │ 4. Dashboard UI Tests (deploy/docker/tests/cli/dashboard/) │ │
│ │ │ │
│ │ - test_01_dashboard_ui.py - Playwright test with screenshots │ │
│ │ - Start server with 3 replicas │ │
│ │ - Run demo_monitor_dashboard.py script │ │
│ │ - Use Playwright to: │ │
│ │ - Take screenshot of main dashboard │ │
│ │ - Verify container filter buttons (All, C-1, C-2, C-3) │ │
│ │ - Test WebSocket connection indicator │ │
│ │ - Verify timeline charts render │ │
│ │ - Test filtering functionality │ │
│ │ - Check all tabs (Requests, Browsers, Janitor, Errors, Stats) │ │
│ │ │ │
│ │ 5. Edge Cases (deploy/docker/tests/cli/edge/) │ │
│ │ │ │
│ │ - test_01_already_running.sh - Try starting when already running │ │
│ │ - test_02_not_running.sh - Try stop/status when not running │ │
│ │ - test_03_scale_single_mode.sh - Try scaling single container mode │ │
│ │ - test_04_invalid_port.sh - Invalid port numbers (0, -1, 99999) │ │
│ │ - test_05_invalid_replicas.sh - Invalid replica counts (0, -1, 101) │ │
│ │ - test_06_missing_env_file.sh - Non-existent env file │ │
│ │ - test_07_port_in_use.sh - Port already occupied │ │
│ │ - test_08_state_corruption.sh - Manually corrupt state file │ │
│ │ - test_09_network_conflict.sh - Docker network name collision │ │
│ │ - test_10_rapid_operations.sh - Start/stop/restart in quick succession │ │
│ │ │ │
│ │ Test Execution Plan │ │
│ │ │ │
│ │ Process: │ │
│ │ │ │
│ │ 1. Create test file │ │
│ │ 2. Run test │ │
│ │ 3. Verify results │ │
│ │ 4. If fails → fix issue → re-test │ │
│ │ 5. Move to next test │ │
│ │ 6. Clean up after each test to ensure clean state │ │
│ │ │ │
│ │ Common Test Structure: │ │
│ │ │ │
│ │ #!/bin/bash │ │
│ │ # Test: [Description] │ │
│ │ # Expected: [What should happen] │ │
│ │ │ │
│ │ source venv/bin/activate │ │
│ │ set -e # Exit on error │ │
│ │ │ │
│ │ echo "=== Test: [Name] ===" │ │
│ │ │ │
│ │ # Setup │ │
│ │ # ... test commands ... │ │
│ │ │ │
│ │ # Verification │ │
│ │ # ... assertions ... │ │
│ │ │ │
│ │ # Cleanup │ │
│ │ crwl server stop || true │ │
│ │ │ │
│ │ echo "✓ Test passed" │ │
│ │ │ │
│ │ Dashboard Test Structure (Python): │ │
│ │ │ │
│ │ # Activate venv first in calling script │ │
│ │ import asyncio │ │
│ │ from playwright.async_api import async_playwright │ │
│ │ │ │
│ │ async def test_dashboard(): │ │
│ │ # Start server with 3 replicas │ │
│ │ # Run demo script in background │ │
│ │ # Launch Playwright │ │
│ │ # Take screenshots │ │
│ │ # Verify elements │ │
│ │ # Cleanup │ │
│ │ │ │
│ │ Success Criteria: │ │
│ │ │ │
│ │ - All basic operations work correctly │ │
│ │ - Scaling operations function properly │ │
│ │ - Resource limits are respected │ │
│ │ - Dashboard UI is functional and responsive │ │
│ │ - Edge cases handled gracefully with proper error messages │ │
│ │ - Clean resource cleanup verified

View File

@@ -1,63 +0,0 @@
#!/bin/bash
# Test: Monitor memory usage during crawl operations
# Expected: Memory stats are accessible and reasonable
set -e
echo "=== Test: Memory Monitoring ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start server
echo "Starting server..."
crwl server start >/dev/null 2>&1
sleep 5
# Get baseline memory
echo "Checking baseline memory..."
BASELINE=$(curl -s http://localhost:11235/monitor/health | jq -r '.container.memory_percent' 2>/dev/null || echo "0")
echo "Baseline memory: ${BASELINE}%"
# Make several crawl requests
echo ""
echo "Making crawl requests to increase memory usage..."
for i in {1..5}; do
echo " Request $i/5..."
curl -s -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d "{\"urls\": [\"https://httpbin.org/html?req=$i\"], \"crawler_config\": {}}" > /dev/null || true
sleep 1
done
# Check memory after requests
echo ""
echo "Checking memory after requests..."
AFTER=$(curl -s http://localhost:11235/monitor/health | jq -r '.container.memory_percent' 2>/dev/null || echo "0")
echo "Memory after requests: ${AFTER}%"
# Get browser pool stats
echo ""
echo "Browser pool memory usage..."
POOL_MEM=$(curl -s http://localhost:11235/monitor/browsers | jq -r '.summary.total_memory_mb' 2>/dev/null || echo "0")
echo "Browser pool: ${POOL_MEM} MB"
# Verify memory is within reasonable bounds (<80%)
MEMORY_OK=$(echo "$AFTER < 80" | bc -l 2>/dev/null || echo "1")
if [[ "$MEMORY_OK" != "1" ]]; then
echo "⚠️ Warning: Memory usage is high: ${AFTER}%"
fi
# Cleanup
echo ""
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Memory monitoring functional"
echo " Baseline: ${BASELINE}%, After: ${AFTER}%, Pool: ${POOL_MEM} MB"

View File

@@ -1,61 +0,0 @@
#!/bin/bash
# Test: CPU usage under concurrent load
# Expected: Server handles concurrent requests without errors
set -e
echo "=== Test: CPU Stress Test ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start server with 3 replicas for better load distribution
echo "Starting server with 3 replicas..."
crwl server start --replicas 3 >/dev/null 2>&1
sleep 12
# Get baseline CPU
echo "Checking baseline container stats..."
docker stats --no-stream --format "table {{.Name}}\t{{.CPUPerc}}\t{{.MemUsage}}" \
--filter "name=crawl4ai" 2>/dev/null || echo "Unable to get container stats"
# Send concurrent requests
echo ""
echo "Sending 10 concurrent requests..."
for i in {1..10}; do
curl -s -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d "{\"urls\": [\"https://httpbin.org/html?req=$i\"], \"crawler_config\": {}}" > /dev/null &
done
# Wait for all requests to complete
echo "Waiting for requests to complete..."
wait
# Check stats after load
echo ""
echo "Container stats after load:"
docker stats --no-stream --format "table {{.Name}}\t{{.CPUPerc}}\t{{.MemUsage}}" \
--filter "name=crawl4ai" 2>/dev/null || echo "Unable to get container stats"
# Verify health
echo ""
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after CPU stress"
crwl server stop
exit 1
fi
# Cleanup
echo ""
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Server handled concurrent load successfully"

View File

@@ -1,72 +0,0 @@
#!/bin/bash
# Test: Start with maximum replicas and stress test
# Expected: Server handles max replicas (10) and distributes load
set -e
echo "=== Test: Maximum Replicas Stress Test ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start with 10 replicas (max recommended)
echo "Starting server with 10 replicas..."
echo "This may take some time..."
crwl server start --replicas 10 >/dev/null 2>&1
sleep 20
# Verify status
echo "Checking status..."
STATUS=$(crwl server status)
if ! echo "$STATUS" | grep -q "10"; then
echo "❌ Failed to start 10 replicas"
crwl server stop
exit 1
fi
# Wait for container discovery
echo ""
echo "Waiting for container discovery..."
sleep 10
# Check containers
CONTAINER_COUNT=$(curl -s http://localhost:11235/monitor/containers | jq -r '.count' 2>/dev/null || echo "0")
echo "Discovered containers: $CONTAINER_COUNT"
# Send burst of requests
echo ""
echo "Sending burst of 20 requests..."
for i in {1..20}; do
curl -s -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d "{\"urls\": [\"https://httpbin.org/html?req=$i\"], \"crawler_config\": {}}" > /dev/null &
done
wait
# Check health after stress
echo ""
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo "❌ Health check failed after max replica stress"
crwl server stop
exit 1
fi
# Check endpoint stats
echo ""
echo "Endpoint statistics:"
curl -s http://localhost:11235/monitor/endpoints/stats | jq '.' 2>/dev/null || echo "No stats available"
# Cleanup
echo ""
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
echo ""
echo "✅ Test passed: Successfully stress tested with 10 replicas"

View File

@@ -1,63 +0,0 @@
#!/bin/bash
# Test: Verify complete resource cleanup
# Expected: All Docker resources are properly removed
set -e
echo "=== Test: Resource Cleanup Verification ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Start server to create resources
echo "Starting server with 3 replicas..."
crwl server start --replicas 3 >/dev/null 2>&1
sleep 10
# List resources before cleanup
echo ""
echo "Resources before cleanup:"
echo "Containers:"
docker ps --filter "name=crawl4ai" --format " - {{.Names}}" 2>/dev/null || echo " None"
docker ps --filter "name=nginx" --format " - {{.Names}}" 2>/dev/null || echo " None"
docker ps --filter "name=redis" --format " - {{.Names}}" 2>/dev/null || echo " None"
echo ""
echo "Networks:"
docker network ls --filter "name=crawl4ai" --format " - {{.Name}}" 2>/dev/null || echo " None"
# Cleanup
echo ""
echo "Performing cleanup..."
crwl server cleanup --force >/dev/null 2>&1
sleep 5
# Verify cleanup
echo ""
echo "Verifying cleanup..."
CONTAINERS=$(docker ps -a --filter "name=crawl4ai" --format "{{.Names}}" 2>/dev/null || echo "")
if [[ -n "$CONTAINERS" ]]; then
echo "❌ Found remaining crawl4ai containers: $CONTAINERS"
exit 1
fi
NGINX=$(docker ps -a --filter "name=nginx" --format "{{.Names}}" 2>/dev/null || echo "")
if [[ -n "$NGINX" ]]; then
echo "⚠️ Warning: Nginx container still exists: $NGINX"
fi
REDIS=$(docker ps -a --filter "name=redis" --format "{{.Names}}" 2>/dev/null || echo "")
if [[ -n "$REDIS" ]]; then
echo "⚠️ Warning: Redis container still exists: $REDIS"
fi
# Verify port is free
if curl -s http://localhost:11235/health > /dev/null 2>&1; then
echo "❌ Port 11235 still in use after cleanup"
exit 1
fi
echo ""
echo "✅ Test passed: All Crawl4AI resources properly cleaned up"

View File

@@ -1,99 +0,0 @@
#!/bin/bash
# Test: Long-running stability test (5 minutes)
# Expected: Server remains stable over extended period
set -e
echo "=== Test: Long-Running Stability (5 minutes) ==="
echo ""
PROJECT_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../../../../" && pwd)"
source "$PROJECT_ROOT/venv/bin/activate"
# Cleanup
crwl server stop 2>/dev/null || true
sleep 2
# Start server
echo "Starting server with 2 replicas..."
crwl server start --replicas 2 >/dev/null 2>&1
sleep 10
# Get start time
START_TIME=$(date +%s)
DURATION=300 # 5 minutes in seconds
REQUEST_COUNT=0
ERROR_COUNT=0
echo ""
echo "Running stability test for 5 minutes..."
echo "Making periodic requests every 10 seconds..."
echo ""
while true; do
CURRENT_TIME=$(date +%s)
ELAPSED=$((CURRENT_TIME - START_TIME))
if [[ $ELAPSED -ge $DURATION ]]; then
break
fi
REMAINING=$((DURATION - ELAPSED))
echo "[$ELAPSED/$DURATION seconds] Remaining: ${REMAINING}s, Requests: $REQUEST_COUNT, Errors: $ERROR_COUNT"
# Make a request
if curl -s -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d '{"urls": ["https://httpbin.org/html"], "crawler_config": {}}' > /dev/null 2>&1; then
REQUEST_COUNT=$((REQUEST_COUNT + 1))
else
ERROR_COUNT=$((ERROR_COUNT + 1))
echo " ⚠️ Request failed"
fi
# Check health every 30 seconds
if [[ $((ELAPSED % 30)) -eq 0 ]]; then
HEALTH=$(curl -s http://localhost:11235/health | jq -r '.status' 2>/dev/null || echo "error")
if [[ "$HEALTH" != "ok" ]]; then
echo " ❌ Health check failed!"
ERROR_COUNT=$((ERROR_COUNT + 1))
fi
# Get memory stats
MEM=$(curl -s http://localhost:11235/monitor/health | jq -r '.container.memory_percent' 2>/dev/null || echo "N/A")
echo " Memory: ${MEM}%"
fi
sleep 10
done
echo ""
echo "Test duration completed!"
echo "Total requests: $REQUEST_COUNT"
echo "Total errors: $ERROR_COUNT"
# Get final stats
echo ""
echo "Final statistics:"
curl -s http://localhost:11235/monitor/endpoints/stats | jq '.' 2>/dev/null || echo "No stats available"
# Verify error rate is acceptable (<10%)
ERROR_RATE=$(echo "scale=2; $ERROR_COUNT * 100 / $REQUEST_COUNT" | bc -l 2>/dev/null || echo "0")
echo ""
echo "Error rate: ${ERROR_RATE}%"
# Cleanup
echo ""
echo "Cleaning up..."
crwl server stop >/dev/null 2>&1
# Check error rate
ERROR_OK=$(echo "$ERROR_RATE < 10" | bc -l 2>/dev/null || echo "1")
if [[ "$ERROR_OK" != "1" ]]; then
echo "❌ Error rate too high: ${ERROR_RATE}%"
exit 1
fi
echo ""
echo "✅ Test passed: Server remained stable over 5 minutes"
echo " Requests: $REQUEST_COUNT, Errors: $ERROR_COUNT, Error rate: ${ERROR_RATE}%"

View File

@@ -1,200 +0,0 @@
#!/bin/bash
# Master Test Runner for Crawl4AI CLI E2E Tests
# Usage: ./run_tests.sh [category] [test_number]
# category: basic|advanced|resource|dashboard|edge|all (default: all)
# test_number: specific test number to run (optional)
set -e
# Color codes for output
RED='\033[0;31m'
GREEN='\033[0;32m'
YELLOW='\033[1;33m'
BLUE='\033[0;34m'
NC='\033[0m' # No Color
# Test counters
TOTAL_TESTS=0
PASSED_TESTS=0
FAILED_TESTS=0
SKIPPED_TESTS=0
# Get script directory
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
# Print header
print_header() {
echo ""
echo "=========================================="
echo "$1"
echo "=========================================="
echo ""
}
# Print test result
print_result() {
local test_name=$1
local result=$2
if [[ "$result" == "PASS" ]]; then
echo -e "${GREEN}✅ PASS${NC}: $test_name"
PASSED_TESTS=$((PASSED_TESTS + 1))
elif [[ "$result" == "FAIL" ]]; then
echo -e "${RED}❌ FAIL${NC}: $test_name"
FAILED_TESTS=$((FAILED_TESTS + 1))
elif [[ "$result" == "SKIP" ]]; then
echo -e "${YELLOW}⏭️ SKIP${NC}: $test_name"
SKIPPED_TESTS=$((SKIPPED_TESTS + 1))
fi
}
# Run a single test
run_test() {
local test_path=$1
local test_name=$(basename "$test_path")
echo ""
echo -e "${BLUE}Running:${NC} $test_name"
echo "----------------------------------------"
TOTAL_TESTS=$((TOTAL_TESTS + 1))
if bash "$test_path"; then
print_result "$test_name" "PASS"
return 0
else
print_result "$test_name" "FAIL"
return 1
fi
}
# Run Python test
run_python_test() {
local test_path=$1
local test_name=$(basename "$test_path")
echo ""
echo -e "${BLUE}Running:${NC} $test_name"
echo "----------------------------------------"
TOTAL_TESTS=$((TOTAL_TESTS + 1))
if python "$test_path"; then
print_result "$test_name" "PASS"
return 0
else
print_result "$test_name" "FAIL"
return 1
fi
}
# Run tests in a category
run_category() {
local category=$1
local test_number=$2
local category_dir="$SCRIPT_DIR/$category"
if [[ ! -d "$category_dir" ]]; then
echo -e "${RED}Error:${NC} Category '$category' not found"
return 1
fi
print_header "Running $category tests"
if [[ -n "$test_number" ]]; then
# Run specific test
local test_file=$(find "$category_dir" -name "*${test_number}*.sh" | head -n 1)
if [[ -z "$test_file" ]]; then
echo -e "${RED}Error:${NC} Test $test_number not found in $category"
return 1
fi
run_test "$test_file"
else
# Run all tests in category
if [[ "$category" == "dashboard" ]]; then
# Dashboard tests are Python
for test_file in "$category_dir"/*.py; do
[[ -f "$test_file" ]] || continue
run_python_test "$test_file" || true
done
else
# Shell script tests
for test_file in "$category_dir"/*.sh; do
[[ -f "$test_file" ]] || continue
run_test "$test_file" || true
done
fi
fi
}
# Print summary
print_summary() {
echo ""
echo "=========================================="
echo "Test Summary"
echo "=========================================="
echo -e "Total: $TOTAL_TESTS"
echo -e "${GREEN}Passed: $PASSED_TESTS${NC}"
echo -e "${RED}Failed: $FAILED_TESTS${NC}"
echo -e "${YELLOW}Skipped: $SKIPPED_TESTS${NC}"
echo ""
if [[ $FAILED_TESTS -eq 0 ]]; then
echo -e "${GREEN}✅ All tests passed!${NC}"
return 0
else
echo -e "${RED}❌ Some tests failed${NC}"
return 1
fi
}
# Main execution
main() {
local category=${1:-all}
local test_number=$2
# Activate virtual environment
if [[ -f "venv/bin/activate" ]]; then
source venv/bin/activate
else
echo -e "${YELLOW}Warning:${NC} venv not found, some tests may fail"
fi
print_header "Crawl4AI CLI E2E Test Suite"
if [[ "$category" == "all" ]]; then
# Run all categories
for cat in basic advanced resource edge; do
run_category "$cat" || true
done
# Dashboard tests separately (can be slow)
echo ""
echo -e "${YELLOW}Note:${NC} Dashboard tests can be run separately with: ./run_tests.sh dashboard"
else
run_category "$category" "$test_number"
fi
print_summary
}
# Show usage
if [[ "$1" == "-h" || "$1" == "--help" ]]; then
echo "Usage: $0 [category] [test_number]"
echo ""
echo "Categories:"
echo " basic - Basic CLI operations (8 tests)"
echo " advanced - Advanced features (8 tests)"
echo " resource - Resource monitoring and stress tests (5 tests)"
echo " dashboard - Dashboard UI tests with Playwright (1 test)"
echo " edge - Edge cases and error handling (10 tests)"
echo " all - Run all tests except dashboard (default)"
echo ""
echo "Examples:"
echo " $0 # Run all tests"
echo " $0 basic # Run all basic tests"
echo " $0 basic 01 # Run test_01 from basic"
echo " $0 dashboard # Run dashboard UI test"
exit 0
fi
main "$@"

View File

@@ -203,51 +203,4 @@ def get_container_memory_percent() -> float:
except:
# Non-container or unsupported: fallback to host
import psutil
return psutil.virtual_memory().percent
def get_container_id() -> str:
"""Get current container ID (hostname in Docker)."""
import socket
return socket.gethostname()
def detect_deployment_mode() -> tuple[str, list[dict]]:
"""Detect if running in single/swarm/compose mode and get container list.
Returns:
(mode, containers) where mode is "single"|"swarm"|"compose"
containers is list of {id, hostname, healthy}
"""
import socket
my_hostname = socket.gethostname()
# Check if we're behind nginx (Compose mode indicator)
# In Compose, service name resolves to multiple IPs
try:
import socket as sock
# Try to resolve "crawl4ai" service name (Compose service)
try:
addrs = sock.getaddrinfo("crawl4ai", None)
unique_ips = set(addr[4][0] for addr in addrs)
if len(unique_ips) > 1:
# Multiple IPs = Compose with replicas
containers = [
{"id": f"container-{i+1}", "hostname": f"crawl4ai-{i+1}", "healthy": True}
for i in range(len(unique_ips))
]
return "compose", containers
except:
pass
# Check for Swarm mode (TODO: needs swarm-specific detection)
# For now, if hostname pattern matches swarm, detect it
if "." in my_hostname and len(my_hostname.split(".")) > 2:
# Swarm hostname format: service.slot.task_id
return "swarm", [{"id": my_hostname, "hostname": my_hostname, "healthy": True}]
except:
pass
# Default: single container
return "single", [{"id": my_hostname, "hostname": my_hostname, "healthy": True}]
return psutil.virtual_memory().percent

View File

@@ -1,147 +0,0 @@
# Crawl4AI cnode - Quick Start Cheat Sheet
Fast reference for getting started with cnode.
---
## 📥 Install
```bash
# Install cnode
curl -sSL https://raw.githubusercontent.com/unclecode/crawl4ai/main/deploy/installer/install-cnode.sh | bash
```
---
## 🚀 Launch Cluster
```bash
# Single server (development)
cnode start
# Production cluster with 5 replicas
cnode start --replicas 5
# Custom port
cnode start --replicas 3 --port 8080
```
---
## 📊 Check Status
```bash
# View server status
cnode status
# View logs
cnode logs -f
```
---
## ⚙️ Scale Cluster
```bash
# Scale to 10 replicas (live, no downtime)
cnode scale 10
# Scale down to 2
cnode scale 2
```
---
## 🔄 Restart/Stop
```bash
# Restart server
cnode restart
# Stop server
cnode stop
```
---
## 🌐 Test the API
```bash
# Simple test - crawl example.com
curl -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d '{
"urls": ["https://example.com"],
"priority": 10
}'
# Pretty print with jq
curl -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d '{
"urls": ["https://example.com"],
"priority": 10
}' | jq '.result.markdown' -r
# Health check
curl http://localhost:11235/health
```
---
## 📱 Monitor Dashboard
```bash
# Open in browser
open http://localhost:11235/monitor
# Or playground
open http://localhost:11235/playground
```
---
## 🐍 Python Example
```python
import requests
response = requests.post(
"http://localhost:11235/crawl",
json={
"urls": ["https://example.com"],
"priority": 10
}
)
result = response.json()
print(result['result']['markdown'])
```
---
## 🎯 Common Commands
| Command | Description |
|---------|-------------|
| `cnode start` | Start server |
| `cnode start -r 5` | Start with 5 replicas |
| `cnode status` | Check status |
| `cnode scale 10` | Scale to 10 replicas |
| `cnode logs -f` | Follow logs |
| `cnode restart` | Restart server |
| `cnode stop` | Stop server |
| `cnode --help` | Show all commands |
---
## 📚 Full Documentation
- **User Guide:** `deploy/installer/USER_GUIDE.md`
- **Developer Docs:** `deploy/installer/README.md`
- **Docker Guide:** `deploy/docker/README.md`
- **Agent Context:** `deploy/docker/AGENT.md`
---
**That's it!** You're ready to crawl at scale 🚀

View File

@@ -1,345 +0,0 @@
# Crawl4AI Node Manager (cnode) - Installation & Distribution
This directory contains the standalone `cnode` package and installation scripts for managing Crawl4AI Docker server instances.
## Overview
`cnode` is a fast, lightweight CLI tool for managing Crawl4AI Docker servers. It provides:
- One-command deployment with automatic scaling
- Single container for development (N=1)
- Docker Swarm for production with built-in load balancing (N>1)
- Docker Compose + Nginx as fallback (N>1)
## Directory Structure
```
deploy/installer/
├── README.md # This file
├── cnode_pkg/ # Standalone Python package
│ ├── __init__.py # Package marker
│ ├── cli.py # CLI interface (commands)
│ ├── server_manager.py # Docker orchestration logic
│ └── requirements.txt # Python dependencies
├── install-cnode.sh # Local installation script
├── install-cnode.sh # Remote installation script (for users)
└── releases/ # Release artifacts for distribution
```
## Installation
### For Users (Remote Installation)
Users can install `cnode` directly from the web:
```bash
# Install from GitHub/website
curl -sSL https://crawl4ai.com/install-cnode.sh | bash
# Or with wget
wget -qO- https://crawl4ai.com/install-cnode.sh | bash
```
### For Local Testing
Test the installation locally:
```bash
cd deploy/installer
./install-cnode.sh
```
## Package Contents
### `cnode_pkg/` - Python Package
This is a self-contained Python package with:
- **`cli.py`**: Click-based CLI with all commands (start, stop, status, scale, logs, cleanup, restart)
- **`server_manager.py`**: Core Docker orchestration logic
- **`requirements.txt`**: Dependencies (click, rich, anyio, pyyaml)
- **`__init__.py`**: Package initialization
### Installation Script
**`install-cnode.sh`** does the following:
1. Checks for Python 3.8+ and pip
2. Checks for Docker (warns if not found)
3. Installs Python dependencies
4. Copies `cnode_pkg/` to `/usr/local/lib/cnode/`
5. Creates wrapper script at `/usr/local/bin/cnode`
6. Verifies installation
### Wrapper Script
Created at `/usr/local/bin/cnode`:
```bash
#!/usr/bin/env bash
set -e
# Find Python
if command -v python3 &> /dev/null; then
PYTHON_CMD="python3"
elif command -v python &> /dev/null; then
PYTHON_CMD="python"
else
echo "Error: Python 3.8+ required" >&2
exit 1
fi
# Run cnode
export PYTHONPATH="/usr/local/lib/cnode:$PYTHONPATH"
exec $PYTHON_CMD -m cnode_pkg.cli "$@"
```
## Performance
**Blazing Fast Startup:**
- **~0.1 seconds** to launch
- 49x faster than compiled binary alternatives
- Minimal overhead, maximum responsiveness
## Requirements
### User Requirements
- Python 3.8 or higher
- pip (Python package manager)
- Docker (for running servers)
### Dependencies (Auto-installed)
- click >= 8.0.0 (CLI framework)
- rich >= 13.0.0 (Terminal formatting)
- anyio >= 3.0.0 (Async I/O)
- pyyaml >= 6.0.0 (YAML parsing)
## Usage
After installation:
```bash
# Quick start
cnode start # Single container on port 11235
cnode start --replicas 5 # 5-replica cluster
cnode status # Check server status
cnode logs -f # Follow logs
cnode scale 10 # Scale to 10 replicas
cnode stop # Stop server
# Get help
cnode --help
cnode start --help
```
## Development Workflow
### Making Changes
1. **Edit source code** in `deploy/docker/`:
```bash
vim deploy/docker/cnode_cli.py
vim deploy/docker/server_manager.py
```
2. **Update package** by copying to installer:
```bash
# Copy CLI
cp deploy/docker/cnode_cli.py deploy/installer/cnode_pkg/cli.py
# Fix imports (deploy.docker → cnode_pkg)
sed -i 's/from deploy\.docker\./from cnode_pkg./g' deploy/installer/cnode_pkg/cli.py
# Copy server manager
cp deploy/docker/server_manager.py deploy/installer/cnode_pkg/server_manager.py
```
3. **Test locally**:
```bash
cd deploy/installer
./install-cnode.sh
cnode --help
```
4. **Commit both**:
```bash
git add deploy/docker/cnode_cli.py
git add deploy/installer/cnode_pkg/cli.py
git commit -m "Update cnode: [description]"
```
### Creating a Release
1. **Tag the release**:
```bash
git tag -a v1.0.0 -m "Release v1.0.0"
git push origin v1.0.0
```
2. **Package for distribution**:
```bash
cd deploy/installer
tar -czf releases/cnode-v1.0.0.tar.gz cnode_pkg/ install-cnode.sh
```
3. **Create GitHub release**:
```bash
gh release create v1.0.0 \
releases/cnode-v1.0.0.tar.gz \
--title "cnode v1.0.0" \
--notes "Release notes here"
```
4. **Update deployment script** (if needed):
- Update `install-cnode.sh` with new version/URL
- Upload to hosting (e.g., `https://crawl4ai.com/install-cnode.sh`)
## Deployment
### Remote Installation Script
The `install-cnode.sh` script is meant to be hosted at a public URL for user installation:
```bash
# Upload to your server
scp install-cnode.sh user@crawl4ai.com:/var/www/html/install-cnode.sh
# Or use GitHub raw URL
https://raw.githubusercontent.com/unclecode/crawl4ai/main/deploy/installer/install-cnode.sh
```
Users can then install with:
```bash
curl -sSL https://crawl4ai.com/install-cnode.sh | bash
```
## Backward Compatibility
The main Crawl4AI CLI (`crwl`) includes a redirect for backward compatibility:
```bash
# These work identically:
crwl server start --replicas 3
cnode start --replicas 3
# All subcommands redirect:
crwl server status → cnode status
crwl server stop → cnode stop
crwl server scale 5 → cnode scale 5
crwl server logs -f → cnode logs -f
```
This ensures existing scripts continue working while users migrate to `cnode`.
## Uninstallation
To remove cnode:
```bash
# Remove command
sudo rm /usr/local/bin/cnode
# Remove package
sudo rm -rf /usr/local/lib/cnode
# (Optional) Uninstall dependencies
pip uninstall click rich anyio pyyaml
```
## Troubleshooting
### Python Not Found
```bash
# Install Python 3.8+
# macOS: brew install python3
# Ubuntu: sudo apt install python3 python3-pip
# RHEL/CentOS: sudo yum install python3 python3-pip
```
### Permission Denied
```bash
# Run installer with sudo
sudo ./install-cnode.sh
# Or change install location
INSTALL_DIR=$HOME/.local/bin ./install-cnode.sh
```
### Command Not Found After Install
```bash
# Add to PATH in ~/.bashrc or ~/.zshrc
export PATH="/usr/local/bin:$PATH"
# Reload shell
source ~/.bashrc # or source ~/.zshrc
```
### Dependencies Install Failed
```bash
# Install manually
pip install --user click rich anyio pyyaml
# Or with break-system-packages (if needed)
pip install --user --break-system-packages click rich anyio pyyaml
```
### Docker Not Running
```bash
# macOS: Start Docker Desktop
# Linux: sudo systemctl start docker
# Check Docker
docker --version
docker ps
```
## Architecture
### Component Flow
```
User runs: cnode start
/usr/local/bin/cnode (wrapper script)
Finds python3 executable
Sets PYTHONPATH=/usr/local/lib/cnode
python3 -m cnode_pkg.cli start
cli.py → start_cmd()
server_manager.py → ServerManager.start()
Docker orchestration (single/swarm/compose)
Server running!
```
### Why Python Wrapper vs Binary?
We chose a Python wrapper over compiled binaries (PyInstaller) because:
| Metric | Python Wrapper | PyInstaller Binary |
|--------|---------------|-------------------|
| Startup time | **0.1s** | 4.7s |
| Size | ~50KB wrapper | 8.8MB |
| Updates | Easy (just copy files) | Rebuild required |
| Dependencies | Python 3.8+ | None |
| Platform | Any with Python | OS-specific builds |
Since users running Crawl4AI already have Python, the wrapper is the clear winner.
## Support
For issues or questions:
- GitHub Issues: https://github.com/unclecode/crawl4ai/issues
- Documentation: https://docs.crawl4ai.com
- Discord: https://discord.gg/crawl4ai
## Version History
- **v1.0.0**: Initial release with Python wrapper approach
- Fast startup (~0.1s)
- Supports single container, Docker Swarm, and Compose modes
- Auto-scaling and load balancing
- Real-time monitoring and logs

View File

@@ -1,676 +0,0 @@
# Crawl4AI Node Manager (cnode) - User Guide 🚀
Self-host your own Crawl4AI server cluster with one command. Scale from development to production effortlessly.
## Table of Contents
- [What is cnode?](#what-is-cnode)
- [Quick Start](#quick-start)
- [Installation](#installation)
- [Basic Usage](#basic-usage)
- [Scaling & Production](#scaling--production)
- [Monitoring Dashboard](#monitoring-dashboard)
- [Using the API](#using-the-api)
- [Management Commands](#management-commands)
- [Troubleshooting](#troubleshooting)
- [Advanced Topics](#advanced-topics)
---
## What is cnode?
**cnode** (Crawl4AI Node Manager) is a CLI tool that manages Crawl4AI Docker server instances with automatic scaling and load balancing.
### Key Features
**One-Command Deployment** - Start a server or cluster instantly
**Automatic Scaling** - Single container or multi-replica cluster
**Built-in Load Balancing** - Docker Swarm or Nginx (auto-detected)
**Real-time Monitoring** - Beautiful web dashboard
**Zero Configuration** - Works out of the box
**Production Ready** - Auto-scaling, health checks, rolling updates
### Architecture Modes
| Replicas | Mode | Load Balancer | Use Case |
|----------|------|---------------|----------|
| 1 | Single Container | None | Development, testing |
| 2+ | Docker Swarm | Built-in | Production (if Swarm available) |
| 2+ | Docker Compose | Nginx | Production (fallback) |
---
## Quick Start
### 1. Install cnode
```bash
# One-line installation
curl -sSL https://crawl4ai.com/install-cnode.sh | bash
```
**Requirements:**
- Python 3.8+
- Docker
- Git
### 2. Start Your First Server
```bash
# Start single development server
cnode start
# Or start a production cluster with 5 replicas
cnode start --replicas 5
```
That's it! Your server is running at **http://localhost:11235** 🎉
---
## Installation
### Method 1: Quick Install (Recommended)
```bash
# From crawl4ai.com (when hosted)
curl -sSL https://crawl4ai.com/install-cnode.sh | bash
# Or directly from GitHub
curl -sSL https://raw.githubusercontent.com/unclecode/crawl4ai/main/deploy/installer/install-cnode.sh | bash
```
### Method 2: Clone Repository (For Development)
```bash
# Clone the repository
git clone https://github.com/unclecode/crawl4ai.git
cd crawl4ai/deploy/installer
# Run installer
./install-cnode.sh
```
### Method 3: Custom Location
```bash
# Install to custom directory (using GitHub raw URL)
INSTALL_DIR=$HOME/.local/bin curl -sSL https://raw.githubusercontent.com/unclecode/crawl4ai/main/deploy/installer/install-cnode.sh | bash
# Add to PATH
export PATH="$HOME/.local/bin:$PATH"
```
### Verify Installation
```bash
cnode --help
```
---
## Basic Usage
### Start Server
```bash
# Development server (1 replica)
cnode start
# Production cluster (5 replicas with auto-scaling)
cnode start --replicas 5
# Custom port
cnode start --port 8080
# Specific Docker image
cnode start --image unclecode/crawl4ai:0.7.0
```
### Check Status
```bash
cnode status
```
**Example Output:**
```
╭─────────────────── Crawl4AI Server Status ───────────────────╮
│ Status │ 🟢 Running │
│ Mode │ swarm │
│ Replicas │ 5 │
│ Port │ 11235 │
│ Image │ unclecode/crawl4ai:latest │
│ Uptime │ 2 hours 34 minutes │
│ Started │ 2025-10-21 14:30:00 │
╰─────────────────────────────────────────────────────────────╯
✓ Server is healthy
Access: http://localhost:11235
```
### View Logs
```bash
# Show last 100 lines
cnode logs
# Follow logs in real-time
cnode logs -f
# Show last 500 lines
cnode logs --tail 500
```
### Stop Server
```bash
# Stop server (keeps data)
cnode stop
# Stop and remove all data
cnode stop --remove-volumes
```
---
## Scaling & Production
### Scale Your Cluster
```bash
# Scale to 10 replicas (live, no downtime)
cnode scale 10
# Scale down to 2 replicas
cnode scale 2
```
**Note:** Scaling is live for Swarm/Compose modes. Single container mode requires restart.
### Production Deployment
```bash
# Start production cluster
cnode start --replicas 5 --port 11235
# Verify health
curl http://localhost:11235/health
# Monitor performance
cnode logs -f
```
### Restart Server
```bash
# Restart with same configuration
cnode restart
# Restart with new replica count
cnode restart --replicas 10
```
---
## Monitoring Dashboard
### Access the Dashboard
Once your server is running, access the real-time monitoring dashboard:
```bash
# Dashboard URL
http://localhost:11235/monitor
```
### Dashboard Features
📊 **Real-time Metrics**
- Requests per second
- Active connections
- Response times
- Error rates
📈 **Performance Graphs**
- CPU usage
- Memory consumption
- Request latency
- Throughput
🔍 **System Health**
- Container status
- Replica health
- Load distribution
- Resource utilization
![Monitor Dashboard](https://crawl4ai.com/images/monitor-dashboard.png)
### API Health Endpoint
```bash
# Quick health check
curl http://localhost:11235/health
# Response
{
"status": "healthy",
"version": "1.0.0",
"uptime": 9876,
"containers": 5
}
```
---
## Using the API
### Interactive Playground
Test the API interactively:
```
http://localhost:11235/playground
```
### Basic Crawl Example
**Python:**
```python
import requests
# Simple crawl
response = requests.post(
"http://localhost:11235/crawl",
json={
"urls": ["https://example.com"],
"browser_config": {
"type": "BrowserConfig",
"params": {"headless": True}
},
"crawler_config": {
"type": "CrawlerRunConfig",
"params": {"cache_mode": "bypass"}
}
}
)
result = response.json()
print(f"Title: {result['result']['metadata']['title']}")
print(f"Content: {result['result']['markdown'][:200]}...")
```
**cURL:**
```bash
curl -X POST http://localhost:11235/crawl \
-H "Content-Type: application/json" \
-d '{
"urls": ["https://example.com"],
"browser_config": {
"type": "BrowserConfig",
"params": {"headless": true}
},
"crawler_config": {
"type": "CrawlerRunConfig",
"params": {"cache_mode": "bypass"}
}
}'
```
**JavaScript (Node.js):**
```javascript
const axios = require('axios');
async function crawl() {
const response = await axios.post('http://localhost:11235/crawl', {
urls: ['https://example.com'],
browser_config: {
type: 'BrowserConfig',
params: { headless: true }
},
crawler_config: {
type: 'CrawlerRunConfig',
params: { cache_mode: 'bypass' }
}
});
console.log('Title:', response.data.result.metadata.title);
console.log('Content:', response.data.result.markdown.substring(0, 200));
}
crawl();
```
### Advanced Examples
**Extract with CSS Selectors:**
```python
import requests
response = requests.post(
"http://localhost:11235/crawl",
json={
"urls": ["https://news.ycombinator.com"],
"browser_config": {
"type": "BrowserConfig",
"params": {"headless": True}
},
"crawler_config": {
"type": "CrawlerRunConfig",
"params": {
"extraction_strategy": {
"type": "JsonCssExtractionStrategy",
"params": {
"schema": {
"type": "dict",
"value": {
"baseSelector": ".athing",
"fields": [
{"name": "title", "selector": ".titleline > a", "type": "text"},
{"name": "url", "selector": ".titleline > a", "type": "attribute", "attribute": "href"},
{"name": "points", "selector": ".score", "type": "text"}
]
}
}
}
}
}
}
}
)
articles = response.json()['result']['extracted_content']
for article in articles:
print(f"{article['title']} - {article['points']}")
```
**Streaming Multiple URLs:**
```python
import requests
import json
response = requests.post(
"http://localhost:11235/crawl/stream",
json={
"urls": [
"https://example.com",
"https://httpbin.org/html",
"https://python.org"
],
"browser_config": {
"type": "BrowserConfig",
"params": {"headless": True}
},
"crawler_config": {
"type": "CrawlerRunConfig",
"params": {"stream": True}
}
},
stream=True
)
for line in response.iter_lines():
if line:
data = json.loads(line)
if data.get("status") == "completed":
break
print(f"Crawled: {data['url']} - Success: {data['success']}")
```
### Additional Endpoints
**Screenshot:**
```bash
curl -X POST http://localhost:11235/screenshot \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com"}' \
--output screenshot.png
```
**PDF Export:**
```bash
curl -X POST http://localhost:11235/pdf \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com"}' \
--output page.pdf
```
**HTML Extraction:**
```bash
curl -X POST http://localhost:11235/html \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com"}'
```
---
## Management Commands
### All Available Commands
```bash
cnode --help # Show help
cnode start [OPTIONS] # Start server
cnode stop [OPTIONS] # Stop server
cnode status # Show status
cnode scale N # Scale to N replicas
cnode logs [OPTIONS] # View logs
cnode restart [OPTIONS] # Restart server
cnode cleanup [--force] # Clean up resources
```
### Command Options
**start:**
```bash
--replicas, -r N # Number of replicas (default: 1)
--mode MODE # Deployment mode: auto, single, swarm, compose
--port, -p PORT # External port (default: 11235)
--env-file FILE # Environment file path
--image IMAGE # Docker image (default: unclecode/crawl4ai:latest)
```
**stop:**
```bash
--remove-volumes # Remove persistent data (WARNING: deletes data)
```
**logs:**
```bash
--follow, -f # Follow log output (like tail -f)
--tail N # Number of lines to show (default: 100)
```
**scale:**
```bash
N # Target replica count (minimum: 1)
```
---
## Troubleshooting
### Server Won't Start
```bash
# Check Docker is running
docker ps
# Check port availability
lsof -i :11235
# Check logs for errors
cnode logs
```
### High Memory Usage
```bash
# Check current status
cnode status
# Restart to clear memory
cnode restart
# Scale down if needed
cnode scale 2
```
### Slow Response Times
```bash
# Scale up for better performance
cnode scale 10
# Check system resources
docker stats
```
### Cannot Connect to API
```bash
# Verify server is running
cnode status
# Check firewall
sudo ufw status
# Test locally
curl http://localhost:11235/health
```
### Clean Slate
```bash
# Complete cleanup and restart
cnode cleanup --force
cnode start --replicas 5
```
---
## Advanced Topics
### Environment Variables
Create `.env` file for API keys:
```bash
# .env file
OPENAI_API_KEY=sk-your-key
ANTHROPIC_API_KEY=your-key
```
Use with cnode:
```bash
cnode start --env-file .env --replicas 3
```
### Custom Docker Image
```bash
# Use specific version
cnode start --image unclecode/crawl4ai:0.7.0-r1
# Use custom registry
cnode start --image myregistry.com/crawl4ai:custom
```
### Production Best Practices
1. **Use Multiple Replicas**
```bash
cnode start --replicas 5
```
2. **Monitor Regularly**
```bash
# Set up monitoring cron
*/5 * * * * cnode status | mail -s "Crawl4AI Status" admin@example.com
```
3. **Regular Log Rotation**
```bash
cnode logs --tail 1000 > crawl4ai.log
cnode restart
```
4. **Resource Limits**
- Ensure adequate RAM (2GB per replica minimum)
- Monitor disk space for cached data
- Use SSD for better performance
### Integration Examples
**With Python App:**
```python
from crawl4ai.docker_client import Crawl4aiDockerClient
async def main():
async with Crawl4aiDockerClient(base_url="http://localhost:11235") as client:
results = await client.crawl(["https://example.com"])
print(results[0].markdown)
```
**With Node.js:**
```javascript
const Crawl4AI = require('crawl4ai-client');
const client = new Crawl4AI('http://localhost:11235');
client.crawl('https://example.com')
.then(result => console.log(result.markdown));
```
**With REST API:**
Any language with HTTP client support can use the API!
---
## Getting Help
### Resources
- 📖 [Full Documentation](https://docs.crawl4ai.com)
- 🐛 [Report Issues](https://github.com/unclecode/crawl4ai/issues)
- 💬 [Discord Community](https://discord.gg/crawl4ai)
- 📺 [Video Tutorials](https://youtube.com/@crawl4ai)
### Common Questions
**Q: How many replicas should I use?**
A: Start with 1 for development. Use 3-5 for production. Scale based on load.
**Q: What's the difference between Swarm and Compose mode?**
A: Swarm has built-in load balancing (faster). Compose uses Nginx (fallback if Swarm unavailable).
**Q: Can I run multiple cnode instances?**
A: Yes! Use different ports: `cnode start --port 8080`
**Q: How do I update to the latest version?**
A: Pull new image: `cnode stop && docker pull unclecode/crawl4ai:latest && cnode start`
---
## Summary
You now know how to:
- ✅ Install cnode with one command
- ✅ Start and manage Crawl4AI servers
- ✅ Scale from 1 to 100+ replicas
- ✅ Monitor performance in real-time
- ✅ Use the API from any language
- ✅ Troubleshoot common issues
**Ready to crawl at scale!** 🚀
For detailed Docker configuration and advanced deployment options, see the [Docker Guide](../docker/README.md).
---
**Happy Crawling!** 🕷️
*Made with ❤️ by the Crawl4AI team*

View File

@@ -1,5 +0,0 @@
"""
Crawl4AI Node Manager (cnode) - Docker server orchestration CLI
"""
__version__ = "1.0.0"

View File

@@ -1,492 +0,0 @@
"""
Crawl4AI Server CLI Commands
Provides `cnode` command group for Docker orchestration.
"""
import click
import anyio
from rich.console import Console
from rich.table import Table
from rich.panel import Panel
from rich.prompt import Confirm
from cnode_pkg.server_manager import ServerManager
console = Console()
@click.group()
def cli():
"""Manage Crawl4AI Docker server instances
\b
One-command deployment with automatic scaling:
• Single container for development (N=1)
• Docker Swarm for production with built-in load balancing (N>1)
• Docker Compose + Nginx as fallback (N>1)
\b
Examples:
cnode start # Single container on port 11235
cnode start --replicas 3 # Auto-detect Swarm or Compose
cnode start -r 5 --port 8080 # 5 replicas on custom port
cnode status # Check current deployment
cnode scale 10 # Scale to 10 replicas
cnode stop # Stop and cleanup
"""
pass
@cli.command("start")
@click.option(
"--replicas", "-r",
type=int,
default=1,
help="Number of container replicas (default: 1)"
)
@click.option(
"--mode",
type=click.Choice(["auto", "single", "swarm", "compose"]),
default="auto",
help="Deployment mode (default: auto-detect)"
)
@click.option(
"--port", "-p",
type=int,
default=11235,
help="External port to expose (default: 11235)"
)
@click.option(
"--env-file",
type=click.Path(exists=True),
help="Path to environment file"
)
@click.option(
"--image",
default="unclecode/crawl4ai:latest",
help="Docker image to use (default: unclecode/crawl4ai:latest)"
)
def start_cmd(replicas: int, mode: str, port: int, env_file: str, image: str):
"""Start Crawl4AI server with automatic orchestration.
Deployment modes:
- auto: Automatically choose best mode (default)
- single: Single container (N=1 only)
- swarm: Docker Swarm with built-in load balancing
- compose: Docker Compose + Nginx reverse proxy
The server will:
1. Check if Docker is running
2. Validate port availability
3. Pull image if needed
4. Start container(s) with health checks
5. Save state for management
Examples:
# Development: single container
cnode start
# Production: 5 replicas with Swarm
cnode start --replicas 5
# Custom configuration
cnode start -r 3 --port 8080 --env-file .env.prod
"""
manager = ServerManager()
console.print(Panel(
f"[cyan]Starting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{replicas}[/yellow]\n"
f"Mode: [yellow]{mode}[/yellow]\n"
f"Port: [yellow]{port}[/yellow]\n"
f"Image: [yellow]{image}[/yellow]",
title="Server Start",
border_style="cyan"
))
with console.status("[cyan]Starting server..."):
async def _start():
return await manager.start(
replicas=replicas,
mode=mode,
port=port,
env_file=env_file,
image=image
)
result = anyio.run(_start)
if result["success"]:
console.print(Panel(
f"[green]✓ Server started successfully![/green]\n\n"
f"Mode: [cyan]{result.get('state_data', {}).get('mode', mode)}[/cyan]\n"
f"URL: [bold]http://localhost:{port}[/bold]\n"
f"Health: [bold]http://localhost:{port}/health[/bold]\n"
f"Monitor: [bold]http://localhost:{port}/monitor[/bold]",
title="Server Running",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Failed to start server[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "already running" in error_msg.lower():
console.print("\n[yellow]Hint: Use 'cnode status' to check current deployment[/yellow]")
console.print("[yellow] Use 'cnode stop' to stop existing server[/yellow]")
@cli.command("status")
def status_cmd():
"""Show current server status and deployment info.
Displays:
- Running state (up/down)
- Deployment mode (single/swarm/compose)
- Number of replicas
- Port mapping
- Uptime
- Image version
Example:
cnode status
"""
manager = ServerManager()
async def _status():
return await manager.status()
result = anyio.run(_status)
if result["running"]:
table = Table(title="Crawl4AI Server Status", border_style="green")
table.add_column("Property", style="cyan")
table.add_column("Value", style="green")
table.add_row("Status", "🟢 Running")
table.add_row("Mode", result["mode"])
table.add_row("Replicas", str(result.get("replicas", 1)))
table.add_row("Port", str(result.get("port", 11235)))
table.add_row("Image", result.get("image", "unknown"))
table.add_row("Uptime", result.get("uptime", "unknown"))
table.add_row("Started", result.get("started_at", "unknown"))
console.print(table)
console.print(f"\n[green]✓ Server is healthy[/green]")
console.print(f"[dim]Access: http://localhost:{result.get('port', 11235)}[/dim]")
else:
console.print(Panel(
f"[yellow]No server is currently running[/yellow]\n\n"
f"Use 'cnode start' to launch a server",
title="Server Status",
border_style="yellow"
))
@cli.command("stop")
@click.option(
"--remove-volumes",
is_flag=True,
help="Remove associated volumes (WARNING: deletes data)"
)
def stop_cmd(remove_volumes: bool):
"""Stop running Crawl4AI server and cleanup resources.
This will:
1. Stop all running containers/services
2. Remove containers
3. Optionally remove volumes (--remove-volumes)
4. Clean up state files
WARNING: Use --remove-volumes with caution as it will delete
persistent data including Redis databases and logs.
Examples:
# Stop server, keep volumes
cnode stop
# Stop and remove all data
cnode stop --remove-volumes
"""
manager = ServerManager()
# Confirm if removing volumes
if remove_volumes:
if not Confirm.ask(
"[red]⚠️ This will delete all server data including Redis databases. Continue?[/red]"
):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Stopping server..."):
async def _stop():
return await manager.stop(remove_volumes=remove_volumes)
result = anyio.run(_stop)
if result["success"]:
console.print(Panel(
f"[green]✓ Server stopped successfully[/green]\n\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Server Stopped",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Error stopping server[/red]\n\n"
f"{result.get('error', result.get('message', 'Unknown error'))}",
title="Error",
border_style="red"
))
@cli.command("scale")
@click.argument("replicas", type=int)
def scale_cmd(replicas: int):
"""Scale server to specified number of replicas.
Only works with Swarm or Compose modes. Single container
mode cannot be scaled (must stop and restart with --replicas).
Scaling is live and does not require downtime. The load
balancer will automatically distribute traffic to new replicas.
Examples:
# Scale up to 10 replicas
cnode scale 10
# Scale down to 2 replicas
cnode scale 2
# Scale to 1 (minimum)
cnode scale 1
"""
if replicas < 1:
console.print("[red]Error: Replicas must be at least 1[/red]")
return
manager = ServerManager()
with console.status(f"[cyan]Scaling to {replicas} replicas..."):
async def _scale():
return await manager.scale(replicas=replicas)
result = anyio.run(_scale)
if result["success"]:
console.print(Panel(
f"[green]✓ Scaled successfully[/green]\n\n"
f"New replica count: [bold]{replicas}[/bold]\n"
f"Mode: [cyan]{result.get('mode')}[/cyan]",
title="Scaling Complete",
border_style="green"
))
else:
error_msg = result.get("error", result.get("message", "Unknown error"))
console.print(Panel(
f"[red]✗ Scaling failed[/red]\n\n"
f"{error_msg}",
title="Error",
border_style="red"
))
if "single container" in error_msg.lower():
console.print("\n[yellow]Hint: For single container mode:[/yellow]")
console.print("[yellow] 1. cnode stop[/yellow]")
console.print(f"[yellow] 2. cnode start --replicas {replicas}[/yellow]")
@cli.command("logs")
@click.option(
"--follow", "-f",
is_flag=True,
help="Follow log output (like tail -f)"
)
@click.option(
"--tail",
type=int,
default=100,
help="Number of lines to show (default: 100)"
)
def logs_cmd(follow: bool, tail: int):
"""View server logs.
Shows logs from running containers/services. Use --follow
to stream logs in real-time.
Examples:
# Show last 100 lines
cnode logs
# Show last 500 lines
cnode logs --tail 500
# Follow logs in real-time
cnode logs --follow
# Combine options
cnode logs -f --tail 50
"""
manager = ServerManager()
async def _logs():
return await manager.logs(follow=follow, tail=tail)
output = anyio.run(_logs)
console.print(output)
@cli.command("cleanup")
@click.option(
"--force",
is_flag=True,
help="Force cleanup even if state file doesn't exist"
)
def cleanup_cmd(force: bool):
"""Force cleanup of all Crawl4AI Docker resources.
Stops and removes all containers, networks, and optionally volumes.
Useful when server is stuck or state is corrupted.
Examples:
# Clean up everything
cnode cleanup
# Force cleanup (ignore state file)
cnode cleanup --force
"""
manager = ServerManager()
console.print(Panel(
f"[yellow]⚠️ Cleaning up Crawl4AI Docker resources[/yellow]\n\n"
f"This will stop and remove:\n"
f"- All Crawl4AI containers\n"
f"- Nginx load balancer\n"
f"- Redis instance\n"
f"- Docker networks\n"
f"- State files",
title="Cleanup",
border_style="yellow"
))
if not force and not Confirm.ask("[yellow]Continue with cleanup?[/yellow]"):
console.print("[yellow]Cancelled[/yellow]")
return
with console.status("[cyan]Cleaning up resources..."):
async def _cleanup():
return await manager.cleanup(force=force)
result = anyio.run(_cleanup)
if result["success"]:
console.print(Panel(
f"[green]✓ Cleanup completed successfully[/green]\n\n"
f"Removed: {result.get('removed', 0)} containers\n"
f"{result.get('message', 'All resources cleaned up')}",
title="Cleanup Complete",
border_style="green"
))
else:
console.print(Panel(
f"[yellow]⚠️ Partial cleanup[/yellow]\n\n"
f"{result.get('message', 'Some resources may still exist')}",
title="Cleanup Status",
border_style="yellow"
))
@cli.command("restart")
@click.option(
"--replicas", "-r",
type=int,
help="New replica count (optional)"
)
def restart_cmd(replicas: int):
"""Restart server (stop then start with same config).
Preserves existing configuration unless overridden with options.
Useful for applying image updates or recovering from errors.
Examples:
# Restart with same configuration
cnode restart
# Restart and change replica count
cnode restart --replicas 5
"""
manager = ServerManager()
# Get current state
async def _get_status():
return await manager.status()
current = anyio.run(_get_status)
if not current["running"]:
console.print("[yellow]No server is running. Use 'cnode start' instead.[/yellow]")
return
# Extract current config
current_replicas = current.get("replicas", 1)
current_port = current.get("port", 11235)
current_image = current.get("image", "unclecode/crawl4ai:latest")
current_mode = current.get("mode", "auto")
# Override with CLI args
new_replicas = replicas if replicas is not None else current_replicas
console.print(Panel(
f"[cyan]Restarting Crawl4AI Server[/cyan]\n\n"
f"Replicas: [yellow]{current_replicas}[/yellow] → [green]{new_replicas}[/green]\n"
f"Port: [yellow]{current_port}[/yellow]\n"
f"Mode: [yellow]{current_mode}[/yellow]",
title="Server Restart",
border_style="cyan"
))
# Stop current
with console.status("[cyan]Stopping current server..."):
async def _stop_server():
return await manager.stop(remove_volumes=False)
stop_result = anyio.run(_stop_server)
if not stop_result["success"]:
console.print(f"[red]Failed to stop server: {stop_result.get('error')}[/red]")
return
# Start new
with console.status("[cyan]Starting server..."):
async def _start_server():
return await manager.start(
replicas=new_replicas,
mode="auto",
port=current_port,
image=current_image
)
start_result = anyio.run(_start_server)
if start_result["success"]:
console.print(Panel(
f"[green]✓ Server restarted successfully![/green]\n\n"
f"URL: [bold]http://localhost:{current_port}[/bold]",
title="Restart Complete",
border_style="green"
))
else:
console.print(Panel(
f"[red]✗ Failed to restart server[/red]\n\n"
f"{start_result.get('error', 'Unknown error')}",
title="Error",
border_style="red"
))
def main():
"""Entry point for cnode CLI"""
cli()
if __name__ == "__main__":
main()
# Test comment

View File

@@ -1,4 +0,0 @@
click>=8.0.0
rich>=13.0.0
anyio>=3.0.0
pyyaml>=6.0.0

File diff suppressed because it is too large Load Diff

View File

@@ -1,171 +0,0 @@
#!/bin/bash
# Crawl4AI Node Manager (cnode) Remote Installation Script
# Usage: curl -sSL https://crawl4ai.com/install-cnode.sh | bash
# Or: wget -qO- https://crawl4ai.com/install-cnode.sh | bash
set -e
# Colors
RED='\033[0;31m'
GREEN='\033[0;32m'
YELLOW='\033[1;33m'
BLUE='\033[0;34m'
NC='\033[0m'
# Configuration
INSTALL_DIR="${INSTALL_DIR:-/usr/local/bin}"
LIB_DIR="${LIB_DIR:-/usr/local/lib/cnode}"
GITHUB_REPO="unclecode/crawl4ai"
BRANCH="${CNODE_BRANCH:-main}"
echo -e "${GREEN}╔══════════════════════════════════════════════════════════════╗${NC}"
echo -e "${GREEN}║ Crawl4AI Node Manager (cnode) Installation Script ║${NC}"
echo -e "${GREEN}╚══════════════════════════════════════════════════════════════╝${NC}\n"
# Check Python
echo -e "${BLUE}Checking Python installation...${NC}"
if command -v python3 &> /dev/null; then
PYTHON_CMD="python3"
elif command -v python &> /dev/null; then
PYTHON_CMD="python"
else
echo -e "${RED}Error: Python 3.8+ is required but not found${NC}"
echo -e "${YELLOW}Install from: https://www.python.org/downloads/${NC}"
exit 1
fi
# Check Python version
PYTHON_VERSION=$($PYTHON_CMD -c 'import sys; print(f"{sys.version_info.major}.{sys.version_info.minor}")')
echo -e "${GREEN}✓ Found Python $PYTHON_VERSION${NC}"
if [ "$(printf '%s\n' "3.8" "$PYTHON_VERSION" | sort -V | head -n1)" != "3.8" ]; then
echo -e "${RED}Error: Python 3.8+ required, found $PYTHON_VERSION${NC}"
exit 1
fi
# Check pip
if ! $PYTHON_CMD -m pip --version &> /dev/null; then
echo -e "${RED}Error: pip is required${NC}"
echo -e "${YELLOW}Install pip: $PYTHON_CMD -m ensurepip${NC}"
exit 1
fi
echo -e "${GREEN}✓ pip is available${NC}"
# Check Docker
echo -e "\n${BLUE}Checking Docker...${NC}"
if ! command -v docker &> /dev/null; then
echo -e "${YELLOW}⚠️ Docker not found (required for running servers)${NC}"
echo -e "${YELLOW}Install from: https://docs.docker.com/get-docker/${NC}\n"
else
echo -e "${GREEN}✓ Docker is installed${NC}"
fi
# Check permissions
USE_SUDO=""
if [ ! -w "$INSTALL_DIR" ] || [ ! -w "/usr/local" ]; then
echo -e "\n${YELLOW}⚠️ Root permission required for installation${NC}"
USE_SUDO="sudo"
fi
# Create temp directory
TMP_DIR="$(mktemp -d)"
cd "$TMP_DIR"
# Download only cnode_pkg from GitHub using sparse checkout
echo -e "\n${BLUE}Downloading cnode package from GitHub...${NC}"
if ! command -v git &> /dev/null; then
echo -e "${RED}Error: git is required but not found${NC}"
echo -e "${YELLOW}Install git and try again${NC}"
rm -rf "$TMP_DIR"
exit 1
fi
# Initialize sparse checkout
git init -q
git remote add origin "https://github.com/$GITHUB_REPO.git"
git config core.sparseCheckout true
# Only checkout the cnode_pkg directory
echo "deploy/installer/cnode_pkg/*" > .git/info/sparse-checkout
# Pull only the needed files
if ! git pull -q --depth=1 origin "$BRANCH"; then
echo -e "${RED}Error: Failed to download package${NC}"
rm -rf "$TMP_DIR"
exit 1
fi
if [ ! -d "deploy/installer/cnode_pkg" ]; then
echo -e "${RED}Error: Package directory not found${NC}"
rm -rf "$TMP_DIR"
exit 1
fi
echo -e "${GREEN}✓ Package downloaded${NC}"
REPO_DIR="."
# Install Python dependencies
echo -e "\n${BLUE}Installing Python dependencies...${NC}"
$PYTHON_CMD -m pip install --quiet --user -r "$REPO_DIR/deploy/installer/cnode_pkg/requirements.txt" 2>/dev/null || \
$PYTHON_CMD -m pip install --quiet --user --break-system-packages -r "$REPO_DIR/deploy/installer/cnode_pkg/requirements.txt" 2>/dev/null || {
echo -e "${YELLOW}⚠️ Could not install dependencies with pip${NC}"
echo -e "${YELLOW}Trying to continue anyway (dependencies may already be installed)${NC}"
}
echo -e "${GREEN}✓ Dependencies check complete${NC}"
# Install cnode package
echo -e "\n${BLUE}Installing cnode package...${NC}"
$USE_SUDO mkdir -p "$LIB_DIR"
$USE_SUDO cp -r "$REPO_DIR/deploy/installer/cnode_pkg" "$LIB_DIR/"
echo -e "${GREEN}✓ Package installed to $LIB_DIR${NC}"
# Create wrapper script
echo -e "\n${BLUE}Creating cnode command...${NC}"
$USE_SUDO tee "$INSTALL_DIR/cnode" > /dev/null << 'EOF'
#!/usr/bin/env bash
# Crawl4AI Node Manager (cnode) wrapper
set -e
# Find Python
if command -v python3 &> /dev/null; then
PYTHON_CMD="python3"
elif command -v python &> /dev/null; then
PYTHON_CMD="python"
else
echo "Error: Python 3.8+ required" >&2
exit 1
fi
# Add cnode to Python path and run
export PYTHONPATH="/usr/local/lib/cnode:$PYTHONPATH"
exec $PYTHON_CMD -m cnode_pkg.cli "$@"
EOF
$USE_SUDO chmod +x "$INSTALL_DIR/cnode"
echo -e "${GREEN}✓ cnode command created${NC}"
# Cleanup
rm -rf "$TMP_DIR"
echo -e "\n${GREEN}✓ Installation complete${NC}"
# Success message
echo -e "\n${GREEN}╔══════════════════════════════════════════════════════════════╗${NC}"
echo -e "${GREEN}║ Installation Complete! ║${NC}"
echo -e "${GREEN}╚══════════════════════════════════════════════════════════════╝${NC}\n"
echo -e "${BLUE}cnode is now installed and ready!${NC}\n"
echo -e "${YELLOW}Quick Start:${NC}"
echo -e " ${GREEN}cnode start${NC} # Start single server"
echo -e " ${GREEN}cnode start --replicas 5${NC} # Start 5-replica cluster"
echo -e " ${GREEN}cnode status${NC} # Check status"
echo -e " ${GREEN}cnode logs -f${NC} # Follow logs"
echo -e " ${GREEN}cnode stop${NC} # Stop server"
echo -e "\n${YELLOW}More help:${NC}"
echo -e " ${BLUE}cnode --help${NC}"
echo -e " ${BLUE}https://github.com/$GITHUB_REPO${NC}\n"

View File

@@ -1,33 +0,0 @@
#!/bin/bash
# Sync cnode source code to installer package
# Run this before committing changes to cnode
set -e
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
SOURCE_DIR="$SCRIPT_DIR/../docker"
PKG_DIR="$SCRIPT_DIR/cnode_pkg"
echo "🔄 Syncing cnode source to package..."
# Copy CLI
echo " → Copying cnode_cli.py to cli.py"
cp "$SOURCE_DIR/cnode_cli.py" "$PKG_DIR/cli.py"
# Fix imports
echo " → Fixing imports (deploy.docker → cnode_pkg)"
sed -i '' 's/from deploy\.docker\./from cnode_pkg./g' "$PKG_DIR/cli.py"
# Copy server manager
echo " → Copying server_manager.py"
cp "$SOURCE_DIR/server_manager.py" "$PKG_DIR/server_manager.py"
echo "✅ Sync complete!"
echo ""
echo "Files updated:"
echo " • deploy/installer/cnode_pkg/cli.py"
echo " • deploy/installer/cnode_pkg/server_manager.py"
echo ""
echo "Next steps:"
echo " 1. Test: cd deploy/installer && ./install-cnode.sh"
echo " 2. Commit both source and package files"

View File

@@ -1,18 +1,44 @@
version: '3.8'
# Shared configuration for all environments
x-base-config: &base-config
ports:
- "11235:11235" # Gunicorn port
env_file:
- .llm.env # API keys (create from .llm.env.example)
# Uncomment to set default environment variables (will overwrite .llm.env)
# environment:
# - OPENAI_API_KEY=${OPENAI_API_KEY:-}
# - DEEPSEEK_API_KEY=${DEEPSEEK_API_KEY:-}
# - ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-}
# - GROQ_API_KEY=${GROQ_API_KEY:-}
# - TOGETHER_API_KEY=${TOGETHER_API_KEY:-}
# - MISTRAL_API_KEY=${MISTRAL_API_KEY:-}
# - GEMINI_API_KEY=${GEMINI_API_KEY:-}
# - LLM_PROVIDER=${LLM_PROVIDER:-} # Optional: Override default provider (e.g., "anthropic/claude-3-opus")
volumes:
- /dev/shm:/dev/shm # Chromium performance
deploy:
resources:
limits:
memory: 4G
reservations:
memory: 1G
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:11235/health"]
interval: 30s
timeout: 10s
retries: 3
start_period: 40s
user: "appuser"
services:
redis:
image: redis:alpine
command: redis-server --appendonly yes
volumes:
- redis_data:/data
networks:
- crawl4ai_net
restart: unless-stopped
crawl4ai:
# 1. Default: Pull multi-platform test image from Docker Hub
# 2. Override with local image via: IMAGE=local-test docker compose up
image: ${IMAGE:-unclecode/crawl4ai:${TAG:-latest}}
# Local build config (used with --build)
build:
context: .
@@ -20,58 +46,6 @@ services:
args:
INSTALL_TYPE: ${INSTALL_TYPE:-default}
ENABLE_GPU: ${ENABLE_GPU:-false}
# No ports exposed - access via nginx only
env_file:
- .llm.env
environment:
- OPENAI_API_KEY=${OPENAI_API_KEY:-}
- DEEPSEEK_API_KEY=${DEEPSEEK_API_KEY:-}
- ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-}
- GROQ_API_KEY=${GROQ_API_KEY:-}
- TOGETHER_API_KEY=${TOGETHER_API_KEY:-}
- MISTRAL_API_KEY=${MISTRAL_API_KEY:-}
- GEMINI_API_TOKEN=${GEMINI_API_TOKEN:-}
- LLM_PROVIDER=${LLM_PROVIDER:-}
- REDIS_HOST=redis
- REDIS_PORT=6379
volumes:
- /dev/shm:/dev/shm # Chromium performance
deploy:
replicas: 3 # Default to 3 replicas (can override with --scale)
resources:
limits:
memory: 4G
reservations:
memory: 1G
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:11235/health"]
interval: 30s
timeout: 10s
retries: 3
start_period: 40s
user: "appuser"
depends_on:
- redis
networks:
- crawl4ai_net
nginx:
image: nginx:alpine
ports:
- "11235:80" # Expose port 11235 to host
volumes:
- ./crawl4ai/templates/nginx.conf.template:/etc/nginx/nginx.conf:ro
depends_on:
- crawl4ai
networks:
- crawl4ai_net
restart: unless-stopped
networks:
crawl4ai_net:
driver: bridge
volumes:
redis_data:
# Inherit shared config
<<: *base-config

626
docs/blog/release-v0.7.7.md Normal file
View File

@@ -0,0 +1,626 @@
# 🚀 Crawl4AI v0.7.7: The Self-Hosting & Monitoring Update
*November 14, 2025 • 10 min read*
---
Today I'm releasing Crawl4AI v0.7.7—the Self-Hosting & Monitoring Update. This release transforms Crawl4AI Docker from a simple containerized crawler into a complete self-hosting platform with enterprise-grade real-time monitoring, full operational transparency, and production-ready observability.
## 🎯 What's New at a Glance
- **📊 Real-time Monitoring Dashboard**: Interactive web UI with live system metrics and browser pool status
- **🔌 Comprehensive Monitor API**: Complete REST API for programmatic access to all monitoring data
- **⚡ WebSocket Streaming**: Real-time updates every 2 seconds for custom dashboards
- **🎮 Control Actions**: Manual browser management (kill, restart, cleanup)
- **🔥 Smart Browser Pool**: 3-tier architecture (permanent/hot/cold) with automatic promotion
- **🧹 Janitor Cleanup System**: Automatic resource management with event logging
- **📈 Production Metrics**: 6 critical metrics for operational excellence
- **🏭 Integration Ready**: Prometheus, alerting, and log aggregation examples
- **🐛 Critical Bug Fixes**: Async LLM extraction, DFS crawling, viewport config, and more
## 📊 Real-time Monitoring Dashboard: Complete Visibility
**The Problem:** Running Crawl4AI in Docker was like flying blind. Users had no visibility into what was happening inside the container—memory usage, active requests, browser pools, or errors. Troubleshooting required checking logs, and there was no way to monitor performance or manually intervene when issues occurred.
**My Solution:** I built a complete real-time monitoring system with an interactive dashboard, comprehensive REST API, WebSocket streaming, and manual control actions. Now you have full transparency and control over your crawling infrastructure.
### The Self-Hosting Value Proposition
Before v0.7.7, Docker was just a containerized crawler. After v0.7.7, it's a complete self-hosting platform that gives you:
- **🔒 Data Privacy**: Your data never leaves your infrastructure
- **💰 Cost Control**: No per-request pricing or rate limits
- **🎯 Full Customization**: Complete control over configurations and strategies
- **📊 Complete Transparency**: Real-time visibility into every aspect
- **⚡ Performance**: Direct access without network overhead
- **🛡️ Enterprise Security**: Keep workflows behind your firewall
### Interactive Monitoring Dashboard
Access the dashboard at `http://localhost:11235/dashboard` to see:
- **System Health Overview**: CPU, memory, network, and uptime in real-time
- **Live Request Tracking**: Active and completed requests with full details
- **Browser Pool Management**: Interactive table with permanent/hot/cold browsers
- **Janitor Events Log**: Automatic cleanup activities
- **Error Monitoring**: Full context error logs
The dashboard updates every 2 seconds via WebSocket, giving you live visibility into your crawling operations.
## 🔌 Monitor API: Programmatic Access
**The Problem:** Monitoring dashboards are great for humans, but automation and integration require programmatic access.
**My Solution:** A comprehensive REST API that exposes all monitoring data for integration with your existing infrastructure.
### System Health Endpoint
```python
import httpx
import asyncio
async def monitor_system_health():
async with httpx.AsyncClient() as client:
response = await client.get("http://localhost:11235/monitor/health")
health = response.json()
print(f"Container Metrics:")
print(f" CPU: {health['container']['cpu_percent']:.1f}%")
print(f" Memory: {health['container']['memory_percent']:.1f}%")
print(f" Uptime: {health['container']['uptime_seconds']}s")
print(f"\nBrowser Pool:")
print(f" Permanent: {health['pool']['permanent']['active']} active")
print(f" Hot Pool: {health['pool']['hot']['count']} browsers")
print(f" Cold Pool: {health['pool']['cold']['count']} browsers")
print(f"\nStatistics:")
print(f" Total Requests: {health['stats']['total_requests']}")
print(f" Success Rate: {health['stats']['success_rate_percent']:.1f}%")
print(f" Avg Latency: {health['stats']['avg_latency_ms']:.0f}ms")
asyncio.run(monitor_system_health())
```
### Request Tracking
```python
async def track_requests():
async with httpx.AsyncClient() as client:
response = await client.get("http://localhost:11235/monitor/requests")
requests_data = response.json()
print(f"Active Requests: {len(requests_data['active'])}")
print(f"Completed Requests: {len(requests_data['completed'])}")
# See details of recent requests
for req in requests_data['completed'][:5]:
status_icon = "" if req['success'] else ""
print(f"{status_icon} {req['endpoint']} - {req['latency_ms']:.0f}ms")
```
### Browser Pool Management
```python
async def monitor_browser_pool():
async with httpx.AsyncClient() as client:
response = await client.get("http://localhost:11235/monitor/browsers")
browsers = response.json()
print(f"Pool Summary:")
print(f" Total Browsers: {browsers['summary']['total_count']}")
print(f" Total Memory: {browsers['summary']['total_memory_mb']} MB")
print(f" Reuse Rate: {browsers['summary']['reuse_rate_percent']:.1f}%")
# List all browsers
for browser in browsers['permanent']:
print(f"🔥 Permanent: {browser['browser_id'][:8]}... | "
f"Requests: {browser['request_count']} | "
f"Memory: {browser['memory_mb']:.0f} MB")
```
### Endpoint Performance Statistics
```python
async def get_endpoint_stats():
async with httpx.AsyncClient() as client:
response = await client.get("http://localhost:11235/monitor/endpoints/stats")
stats = response.json()
print("Endpoint Analytics:")
for endpoint, data in stats.items():
print(f" {endpoint}:")
print(f" Requests: {data['count']}")
print(f" Avg Latency: {data['avg_latency_ms']:.0f}ms")
print(f" Success Rate: {data['success_rate_percent']:.1f}%")
```
### Complete API Reference
The Monitor API includes these endpoints:
- `GET /monitor/health` - System health with pool statistics
- `GET /monitor/requests` - Active and completed request tracking
- `GET /monitor/browsers` - Browser pool details and efficiency
- `GET /monitor/endpoints/stats` - Per-endpoint performance analytics
- `GET /monitor/timeline?minutes=5` - Time-series data for charts
- `GET /monitor/logs/janitor?limit=10` - Cleanup activity logs
- `GET /monitor/logs/errors?limit=10` - Error logs with context
- `POST /monitor/actions/cleanup` - Force immediate cleanup
- `POST /monitor/actions/kill_browser` - Kill specific browser
- `POST /monitor/actions/restart_browser` - Restart browser
- `POST /monitor/stats/reset` - Reset accumulated statistics
## ⚡ WebSocket Streaming: Real-time Updates
**The Problem:** Polling the API every few seconds wastes resources and adds latency. Real-time dashboards need instant updates.
**My Solution:** WebSocket streaming with 2-second update intervals for building custom real-time dashboards.
### WebSocket Integration Example
```python
import websockets
import json
import asyncio
async def monitor_realtime():
uri = "ws://localhost:11235/monitor/ws"
async with websockets.connect(uri) as websocket:
print("Connected to real-time monitoring stream")
while True:
# Receive update every 2 seconds
data = await websocket.recv()
update = json.loads(data)
# Access all monitoring data
print(f"\n--- Update at {update['timestamp']} ---")
print(f"Memory: {update['health']['container']['memory_percent']:.1f}%")
print(f"Active Requests: {len(update['requests']['active'])}")
print(f"Total Browsers: {update['browsers']['summary']['total_count']}")
if update['errors']:
print(f"⚠️ Recent Errors: {len(update['errors'])}")
asyncio.run(monitor_realtime())
```
**Expected Real-World Impact:**
- **Custom Dashboards**: Build tailored monitoring UIs for your team
- **Real-time Alerting**: Trigger alerts instantly when metrics exceed thresholds
- **Integration**: Feed live data into monitoring tools like Grafana
- **Automation**: React to events in real-time without polling
## 🔥 Smart Browser Pool: 3-Tier Architecture
**The Problem:** Creating a new browser for every request is slow and memory-intensive. Traditional browser pools are static and inefficient.
**My Solution:** A smart 3-tier browser pool that automatically adapts to usage patterns.
### How It Works
```python
import httpx
async def demonstrate_browser_pool():
async with httpx.AsyncClient() as client:
# Request 1-3: Default config → Uses permanent browser
print("Phase 1: Using permanent browser")
for i in range(3):
await client.post(
"http://localhost:11235/crawl",
json={"urls": [f"https://httpbin.org/html?req={i}"]}
)
print(f" Request {i+1}: Reused permanent browser")
# Request 4-6: Custom viewport → Cold pool (first use)
print("\nPhase 2: Custom config creates cold pool browser")
viewport_config = {"viewport": {"width": 1280, "height": 720}}
for i in range(4):
await client.post(
"http://localhost:11235/crawl",
json={
"urls": [f"https://httpbin.org/json?v={i}"],
"browser_config": viewport_config
}
)
if i < 2:
print(f" Request {i+1}: Cold pool browser")
else:
print(f" Request {i+1}: Promoted to hot pool! (after 3 uses)")
# Check pool status
response = await client.get("http://localhost:11235/monitor/browsers")
browsers = response.json()
print(f"\nPool Status:")
print(f" Permanent: {len(browsers['permanent'])} (always active)")
print(f" Hot: {len(browsers['hot'])} (frequently used configs)")
print(f" Cold: {len(browsers['cold'])} (on-demand)")
print(f" Reuse Rate: {browsers['summary']['reuse_rate_percent']:.1f}%")
asyncio.run(demonstrate_browser_pool())
```
**Pool Tiers:**
- **🔥 Permanent Browser**: Always-on, default configuration, instant response
- **♨️ Hot Pool**: Browsers promoted after 3+ uses, kept warm for quick access
- **❄️ Cold Pool**: On-demand browsers for variant configs, cleaned up when idle
**Expected Real-World Impact:**
- **Memory Efficiency**: 10x reduction in memory usage vs creating browsers per request
- **Performance**: Instant access to frequently-used configurations
- **Automatic Optimization**: Pool adapts to your usage patterns
- **Resource Management**: Janitor automatically cleans up idle browsers
## 🧹 Janitor System: Automatic Cleanup
**The Problem:** Long-running crawlers accumulate idle browsers and consume memory over time.
**My Solution:** An automatic janitor system that monitors and cleans up idle resources.
```python
async def monitor_janitor_activity():
async with httpx.AsyncClient() as client:
response = await client.get("http://localhost:11235/monitor/logs/janitor?limit=5")
logs = response.json()
print("Recent Cleanup Activities:")
for log in logs:
print(f" {log['timestamp']}: {log['message']}")
# Example output:
# 2025-11-14 10:30:00: Cleaned up 2 cold pool browsers (idle > 5min)
# 2025-11-14 10:25:00: Browser reuse rate: 85.3%
# 2025-11-14 10:20:00: Hot pool browser promoted (10 requests)
```
## 🎮 Control Actions: Manual Management
**The Problem:** Sometimes you need to manually intervene—kill a stuck browser, force cleanup, or restart resources.
**My Solution:** Manual control actions via the API for operational troubleshooting.
### Force Cleanup
```python
async def force_cleanup():
async with httpx.AsyncClient() as client:
response = await client.post("http://localhost:11235/monitor/actions/cleanup")
result = response.json()
print(f"Cleanup completed:")
print(f" Browsers cleaned: {result.get('cleaned_count', 0)}")
print(f" Memory freed: {result.get('memory_freed_mb', 0):.1f} MB")
```
### Kill Specific Browser
```python
async def kill_stuck_browser(browser_id: str):
async with httpx.AsyncClient() as client:
response = await client.post(
"http://localhost:11235/monitor/actions/kill_browser",
json={"browser_id": browser_id}
)
if response.status_code == 200:
print(f"✅ Browser {browser_id} killed successfully")
```
### Reset Statistics
```python
async def reset_stats():
async with httpx.AsyncClient() as client:
response = await client.post("http://localhost:11235/monitor/stats/reset")
print("📊 Statistics reset for fresh monitoring")
```
## 📈 Production Integration Patterns
### Prometheus Integration
```python
# Export metrics for Prometheus scraping
async def export_prometheus_metrics():
async with httpx.AsyncClient() as client:
health = await client.get("http://localhost:11235/monitor/health")
data = health.json()
# Export in Prometheus format
metrics = f"""
# HELP crawl4ai_memory_usage_percent Memory usage percentage
# TYPE crawl4ai_memory_usage_percent gauge
crawl4ai_memory_usage_percent {data['container']['memory_percent']}
# HELP crawl4ai_request_success_rate Request success rate
# TYPE crawl4ai_request_success_rate gauge
crawl4ai_request_success_rate {data['stats']['success_rate_percent']}
# HELP crawl4ai_browser_pool_count Total browsers in pool
# TYPE crawl4ai_browser_pool_count gauge
crawl4ai_browser_pool_count {data['pool']['permanent']['active'] + data['pool']['hot']['count'] + data['pool']['cold']['count']}
"""
return metrics
```
### Alerting Example
```python
async def check_alerts():
async with httpx.AsyncClient() as client:
health = await client.get("http://localhost:11235/monitor/health")
data = health.json()
# Memory alert
if data['container']['memory_percent'] > 80:
print("🚨 ALERT: Memory usage above 80%")
# Trigger cleanup
await client.post("http://localhost:11235/monitor/actions/cleanup")
# Success rate alert
if data['stats']['success_rate_percent'] < 90:
print("🚨 ALERT: Success rate below 90%")
# Check error logs
errors = await client.get("http://localhost:11235/monitor/logs/errors")
print(f"Recent errors: {len(errors.json())}")
# Latency alert
if data['stats']['avg_latency_ms'] > 5000:
print("🚨 ALERT: Average latency above 5s")
```
### Key Metrics to Track
```python
CRITICAL_METRICS = {
"memory_usage": {
"current": "container.memory_percent",
"target": "<80%",
"alert_threshold": ">80%",
"action": "Force cleanup or scale"
},
"success_rate": {
"current": "stats.success_rate_percent",
"target": ">95%",
"alert_threshold": "<90%",
"action": "Check error logs"
},
"avg_latency": {
"current": "stats.avg_latency_ms",
"target": "<2000ms",
"alert_threshold": ">5000ms",
"action": "Investigate slow requests"
},
"browser_reuse_rate": {
"current": "browsers.summary.reuse_rate_percent",
"target": ">80%",
"alert_threshold": "<60%",
"action": "Check pool configuration"
},
"total_browsers": {
"current": "browsers.summary.total_count",
"target": "<15",
"alert_threshold": ">20",
"action": "Check for browser leaks"
},
"error_frequency": {
"current": "len(errors)",
"target": "<5/hour",
"alert_threshold": ">10/hour",
"action": "Review error patterns"
}
}
```
## 🐛 Critical Bug Fixes
This release includes significant bug fixes that improve stability and performance:
### Async LLM Extraction (#1590)
**The Problem:** LLM extraction was blocking async execution, causing URLs to be processed sequentially instead of in parallel (issue #1055).
**The Fix:** Resolved the blocking issue to enable true parallel processing for LLM extraction.
```python
# Before v0.7.7: Sequential processing
# After v0.7.7: True parallel processing
async with AsyncWebCrawler() as crawler:
urls = ["url1", "url2", "url3", "url4"]
# Now processes truly in parallel with LLM extraction
results = await crawler.arun_many(
urls,
config=CrawlerRunConfig(
extraction_strategy=LLMExtractionStrategy(...)
)
)
# 4x faster for parallel LLM extraction!
```
**Expected Impact:** Major performance improvement for batch LLM extraction workflows.
### DFS Deep Crawling (#1607)
**The Problem:** DFS (Depth-First Search) deep crawl strategy had implementation issues.
**The Fix:** Enhanced DFSDeepCrawlStrategy with proper seen URL tracking and improved documentation.
### Browser & Crawler Config Documentation (#1609)
**The Problem:** Documentation didn't match the actual `async_configs.py` implementation.
**The Fix:** Updated all configuration documentation to accurately reflect the current implementation.
### Sitemap Seeder (#1598)
**The Problem:** Sitemap parsing and URL normalization issues in AsyncUrlSeeder (issue #1559).
**The Fix:** Added comprehensive tests and fixes for sitemap namespace parsing and URL normalization.
### Remove Overlay Elements (#1529)
**The Problem:** The `remove_overlay_elements` functionality wasn't working (issue #1396).
**The Fix:** Fixed by properly calling the injected JavaScript function.
### Viewport Configuration (#1495)
**The Problem:** Viewport configuration wasn't working in managed browsers (issue #1490).
**The Fix:** Added proper viewport size configuration support for browser launch.
### Managed Browser CDP Timing (#1528)
**The Problem:** CDP (Chrome DevTools Protocol) endpoint verification had timing issues causing connection failures (issue #1445).
**The Fix:** Added exponential backoff for CDP endpoint verification to handle timing variations.
### Security Updates
- **pyOpenSSL**: Updated from >=24.3.0 to >=25.3.0 to address security vulnerability
- Added verification tests for the security update
### Docker Fixes
- **Port Standardization**: Fixed inconsistent port usage (11234 vs 11235) - now standardized to 11235
- **LLM Environment**: Fixed LLM API key handling for multi-provider support (PR #1537)
- **Error Handling**: Improved Docker API error messages with comprehensive status codes
- **Serialization**: Fixed `fit_html` property serialization in `/crawl` and `/crawl/stream` endpoints
### Other Important Fixes
- **arun_many Returns**: Fixed function to always return a list, even on exception (PR #1530)
- **Webhook Serialization**: Properly serialize Pydantic HttpUrl in webhook config
- **LLMConfig Documentation**: Fixed casing and variable name consistency (issue #1551)
- **Python Version**: Dropped Python 3.9 support, now requires Python >=3.10
## 📊 Expected Real-World Impact
### For DevOps & Infrastructure Teams
- **Full Visibility**: Know exactly what's happening inside your crawling infrastructure
- **Proactive Monitoring**: Catch issues before they become problems
- **Resource Optimization**: Identify memory leaks and performance bottlenecks
- **Operational Control**: Manual intervention when automated systems need help
### For Production Deployments
- **Enterprise Observability**: Prometheus, Grafana, and alerting integration
- **Debugging**: Real-time logs and error tracking
- **Capacity Planning**: Historical metrics for scaling decisions
- **SLA Monitoring**: Track success rates and latency against targets
### For Development Teams
- **Local Monitoring**: Understand crawler behavior during development
- **Performance Testing**: Measure impact of configuration changes
- **Troubleshooting**: Quickly identify and fix issues
- **Learning**: See exactly how the browser pool works
## 🔄 Breaking Changes
**None!** This release is fully backward compatible.
- All existing Docker configurations continue to work
- No API changes to existing endpoints
- Monitoring is additive functionality
- No migration required
## 🚀 Upgrade Instructions
### Docker
```bash
# Pull the latest version
docker pull unclecode/crawl4ai:0.7.7
# Or use the latest tag
docker pull unclecode/crawl4ai:latest
# Run with monitoring enabled (default)
docker run -d \
-p 11235:11235 \
--shm-size=1g \
--name crawl4ai \
unclecode/crawl4ai:0.7.7
# Access the monitoring dashboard
open http://localhost:11235/dashboard
```
### Python Package
```bash
# Upgrade to latest version
pip install --upgrade crawl4ai
# Or install specific version
pip install crawl4ai==0.7.7
```
## 🎬 Try the Demo
Run the comprehensive demo that showcases all monitoring features:
```bash
python docs/releases_review/demo_v0.7.7.py
```
**The demo includes:**
1. System health overview with live metrics
2. Request tracking with active/completed monitoring
3. Browser pool management (permanent/hot/cold)
4. Complete Monitor API endpoint examples
5. WebSocket streaming demonstration
6. Control actions (cleanup, kill, restart)
7. Production metrics and alerting patterns
8. Self-hosting value proposition
## 📚 Documentation
### New Documentation
- **[Self-Hosting Guide](https://docs.crawl4ai.com/core/self-hosting/)** - Complete self-hosting documentation with monitoring
- **Demo Script**: `docs/releases_review/demo_v0.7.7.py` - Working examples
### Updated Documentation
- **Docker Deployment** → **Self-Hosting** (renamed for better positioning)
- Added comprehensive monitoring sections
- Production integration patterns
- WebSocket streaming examples
## 💡 Pro Tips
1. **Start with the dashboard** - Visit `/dashboard` to get familiar with the monitoring system
2. **Track the 6 key metrics** - Memory, success rate, latency, reuse rate, browser count, errors
3. **Set up alerting early** - Use the Monitor API to build alerts before issues occur
4. **Monitor browser pool efficiency** - Aim for >80% reuse rate for optimal performance
5. **Use WebSocket for custom dashboards** - Build tailored monitoring UIs for your team
6. **Leverage Prometheus integration** - Export metrics for long-term storage and analysis
7. **Check janitor logs** - Understand automatic cleanup patterns
8. **Use control actions judiciously** - Manual interventions are for exceptional cases
## 🙏 Acknowledgments
Thank you to our community for the feedback, bug reports, and feature requests that shaped this release. Special thanks to everyone who contributed to the issues that were fixed in this version.
The monitoring system was built based on real user needs for production deployments, and your input made it comprehensive and practical.
## 📞 Support & Resources
- **📖 Documentation**: [docs.crawl4ai.com](https://docs.crawl4ai.com)
- **🐙 GitHub**: [github.com/unclecode/crawl4ai](https://github.com/unclecode/crawl4ai)
- **💬 Discord**: [discord.gg/crawl4ai](https://discord.gg/jP8KfhDhyN)
- **🐦 Twitter**: [@unclecode](https://x.com/unclecode)
- **📊 Dashboard**: `http://localhost:11235/dashboard` (when running)
---
**Crawl4AI v0.7.7 delivers complete self-hosting with enterprise-grade monitoring. You now have full visibility and control over your web crawling infrastructure. The monitoring dashboard, comprehensive API, and WebSocket streaming give you everything needed for production deployments. Try the self-hosting platform—it's a game changer for operational excellence!**
**Happy crawling with full visibility!** 🕷️📊
*- unclecode*

View File

@@ -18,7 +18,7 @@ A comprehensive web-based tutorial for learning and experimenting with C4A-Scrip
2. **Install Dependencies**
```bash
pip install flask
pip install -r requirements.txt
```
3. **Launch the Server**
@@ -28,7 +28,7 @@ A comprehensive web-based tutorial for learning and experimenting with C4A-Scrip
4. **Open in Browser**
```
http://localhost:8080
http://localhost:8000
```
**🌐 Try Online**: [Live Demo](https://docs.crawl4ai.com/c4a-script/demo)
@@ -325,7 +325,7 @@ Powers the recording functionality:
### Configuration
```python
# server.py configuration
PORT = 8080
PORT = 8000
DEBUG = True
THREADED = True
```
@@ -343,9 +343,9 @@ THREADED = True
**Port Already in Use**
```bash
# Kill existing process
lsof -ti:8080 | xargs kill -9
lsof -ti:8000 | xargs kill -9
# Or use different port
python server.py --port 8081
python server.py --port 8001
```
**Blockly Not Loading**

View File

@@ -216,7 +216,7 @@ def get_examples():
'name': 'Handle Cookie Banner',
'description': 'Accept cookies and close newsletter popup',
'script': '''# Handle cookie banner and newsletter
GO http://127.0.0.1:8080/playground/
GO http://127.0.0.1:8000/playground/
WAIT `body` 2
IF (EXISTS `.cookie-banner`) THEN CLICK `.accept`
IF (EXISTS `.newsletter-popup`) THEN CLICK `.close`'''

View File

@@ -0,0 +1,62 @@
import asyncio
import capsolver
from crawl4ai import *
# TODO: set your config
# Docs: https://docs.capsolver.com/guide/captcha/awsWaf/
api_key = "CAP-xxxxxxxxxxxxxxxxxxxxx" # your api key of capsolver
site_url = "https://nft.porsche.com/onboarding@6" # page url of your target site
cookie_domain = ".nft.porsche.com" # the domain name to which you want to apply the cookie
captcha_type = "AntiAwsWafTaskProxyLess" # type of your target captcha
capsolver.api_key = api_key
async def main():
browser_config = BrowserConfig(
verbose=True,
headless=False,
use_persistent_context=True,
)
async with AsyncWebCrawler(config=browser_config) as crawler:
await crawler.arun(
url=site_url,
cache_mode=CacheMode.BYPASS,
session_id="session_captcha_test"
)
# get aws waf cookie using capsolver sdk
solution = capsolver.solve({
"type": captcha_type,
"websiteURL": site_url,
})
cookie = solution["cookie"]
print("aws waf cookie:", cookie)
js_code = """
document.cookie = \'aws-waf-token=""" + cookie + """;domain=""" + cookie_domain + """;path=/\';
location.reload();
"""
wait_condition = """() => {
return document.title === \'Join Porsches journey into Web3\';
}"""
run_config = CrawlerRunConfig(
cache_mode=CacheMode.BYPASS,
session_id="session_captcha_test",
js_code=js_code,
js_only=True,
wait_for=f"js:{wait_condition}"
)
result_next = await crawler.arun(
url=site_url,
config=run_config,
)
print(result_next.markdown)
if __name__ == "__main__":
asyncio.run(main())

View File

@@ -0,0 +1,60 @@
import asyncio
import capsolver
from crawl4ai import *
# TODO: set your config
# Docs: https://docs.capsolver.com/guide/captcha/cloudflare_challenge/
api_key = "CAP-xxxxxxxxxxxxxxxxxxxxx" # your api key of capsolver
site_url = "https://gitlab.com/users/sign_in" # page url of your target site
captcha_type = "AntiCloudflareTask" # type of your target captcha
# your http proxy to solve cloudflare challenge
proxy_server = "proxy.example.com:8080"
proxy_username = "myuser"
proxy_password = "mypass"
capsolver.api_key = api_key
async def main():
# get challenge cookie using capsolver sdk
solution = capsolver.solve({
"type": captcha_type,
"websiteURL": site_url,
"proxy": f"{proxy_server}:{proxy_username}:{proxy_password}",
})
cookies = solution["cookies"]
user_agent = solution["userAgent"]
print("challenge cookies:", cookies)
cookies_list = []
for name, value in cookies.items():
cookies_list.append({
"name": name,
"value": value,
"url": site_url,
})
browser_config = BrowserConfig(
verbose=True,
headless=False,
use_persistent_context=True,
user_agent=user_agent,
cookies=cookies_list,
proxy_config={
"server": f"http://{proxy_server}",
"username": proxy_username,
"password": proxy_password,
},
)
async with AsyncWebCrawler(config=browser_config) as crawler:
result = await crawler.arun(
url=site_url,
cache_mode=CacheMode.BYPASS,
session_id="session_captcha_test"
)
print(result.markdown)
if __name__ == "__main__":
asyncio.run(main())

View File

@@ -0,0 +1,64 @@
import asyncio
import capsolver
from crawl4ai import *
# TODO: set your config
# Docs: https://docs.capsolver.com/guide/captcha/cloudflare_turnstile/
api_key = "CAP-xxxxxxxxxxxxxxxxxxxxx" # your api key of capsolver
site_key = "0x4AAAAAAAGlwMzq_9z6S9Mh" # site key of your target site
site_url = "https://clifford.io/demo/cloudflare-turnstile" # page url of your target site
captcha_type = "AntiTurnstileTaskProxyLess" # type of your target captcha
capsolver.api_key = api_key
async def main():
browser_config = BrowserConfig(
verbose=True,
headless=False,
use_persistent_context=True,
)
async with AsyncWebCrawler(config=browser_config) as crawler:
await crawler.arun(
url=site_url,
cache_mode=CacheMode.BYPASS,
session_id="session_captcha_test"
)
# get turnstile token using capsolver sdk
solution = capsolver.solve({
"type": captcha_type,
"websiteURL": site_url,
"websiteKey": site_key,
})
token = solution["token"]
print("turnstile token:", token)
js_code = """
document.querySelector(\'input[name="cf-turnstile-response"]\').value = \'"""+token+"""\';
document.querySelector(\'button[type="submit"]\').click();
"""
wait_condition = """() => {
const items = document.querySelectorAll(\'h1\');
return items.length === 0;
}"""
run_config = CrawlerRunConfig(
cache_mode=CacheMode.BYPASS,
session_id="session_captcha_test",
js_code=js_code,
js_only=True,
wait_for=f"js:{wait_condition}"
)
result_next = await crawler.arun(
url=site_url,
config=run_config,
)
print(result_next.markdown)
if __name__ == "__main__":
asyncio.run(main())

Some files were not shown because too many files have changed in this diff Show More