From 3efb59fb9a344c3a89e3af8926190151961c2a1e Mon Sep 17 00:00:00 2001 From: ntohidi Date: Wed, 22 Oct 2025 13:14:11 +0200 Subject: [PATCH] fix: update Crawl4AI Docker container port from 11234 to 11235 --- docs/examples/docker_webhook_example.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docs/examples/docker_webhook_example.py b/docs/examples/docker_webhook_example.py index 8822e879..f05d3501 100644 --- a/docs/examples/docker_webhook_example.py +++ b/docs/examples/docker_webhook_example.py @@ -9,7 +9,7 @@ Supports both: - /llm/job - LLM-powered content extraction Prerequisites: -1. Crawl4AI Docker container running on localhost:11234 +1. Crawl4AI Docker container running on localhost:11235 2. Flask installed: pip install flask requests 3. LLM API key configured in .llm.env (for LLM extraction examples) @@ -26,7 +26,7 @@ from flask import Flask, request, jsonify from threading import Thread # Configuration -CRAWL4AI_BASE_URL = "http://localhost:11234" +CRAWL4AI_BASE_URL = "http://localhost:11235" WEBHOOK_BASE_URL = "http://localhost:8080" # Your webhook receiver URL # Initialize Flask app for webhook receiver @@ -344,7 +344,7 @@ def main(): except: print(f"❌ Cannot connect to Crawl4AI at {CRAWL4AI_BASE_URL}") print(" Please make sure Docker container is running:") - print(" docker run -d -p 11234:11234 --name crawl4ai unclecode/crawl4ai:latest") + print(" docker run -d -p 11235:11235 --name crawl4ai unclecode/crawl4ai:latest") return # Start webhook server in background thread