#1268 fix: update redirected_url to current page URL and enhance normalize_url function

This commit is contained in:
AHMET YILMAZ
2025-09-08 19:09:33 +08:00
parent 0482c1eafc
commit 813b1f5534
3 changed files with 53 additions and 10 deletions

View File

@@ -1037,7 +1037,7 @@ class AsyncPlaywrightCrawlerStrategy(AsyncCrawlerStrategy):
downloaded_files=(
self._downloaded_files if self._downloaded_files else None
),
redirected_url=redirected_url,
redirected_url=page.url, # Update to current URL in case of JavaScript navigation
# Include captured data if enabled
network_requests=captured_requests if config.capture_network_requests else None,
console_messages=captured_console if config.capture_console_messages else None,

View File

@@ -480,7 +480,7 @@ class AsyncWebCrawler:
# Scraping Strategy Execution #
################################
result: ScrapingResult = scraping_strategy.scrap(
url, html, **params)
kwargs.get("redirected_url", url), html, **params)
if result is None:
raise ValueError(

View File

@@ -2149,8 +2149,10 @@ def normalize_url(
*,
drop_query_tracking=True,
sort_query=True,
keep_fragment=False,
keep_fragment=True,
remove_fragments=None, # alias for keep_fragment=False
extra_drop_params=None,
params_to_remove=None, # alias for extra_drop_params
preserve_https=False,
original_scheme=None
):
@@ -2175,10 +2177,20 @@ def normalize_url(
Returns
-------
str | None
A clean, canonical URL or None if href is empty/None.
A clean, canonical URL or the base URL if href is empty/None.
"""
if not href:
return None
# For empty href, return the base URL (matching urljoin behavior)
return base_url
# Validate base URL format
parsed_base = urlparse(base_url)
if not parsed_base.scheme or not parsed_base.netloc:
raise ValueError(f"Invalid base URL format: {base_url}")
if parsed_base.scheme.lower() not in ["http", "https"]:
# Handle special protocols
raise ValueError(f"Invalid base URL format: {base_url}")
# Resolve relative paths first
full_url = urljoin(base_url, href.strip())
@@ -2199,6 +2211,12 @@ def normalize_url(
# ── netloc ──
netloc = parsed.netloc.lower()
# Remove default ports (80 for http, 443 for https)
if ':' in netloc:
host, port = netloc.rsplit(':', 1)
if (parsed.scheme == 'http' and port == '80') or (parsed.scheme == 'https' and port == '443'):
netloc = host
# ── path ──
# Strip duplicate slashes and trailing "/" (except root)
@@ -2206,7 +2224,17 @@ def normalize_url(
# The path from urlparse is already properly encoded
path = parsed.path
if path.endswith('/') and path != '/':
path = path.rstrip('/')
# Only strip trailing slash if the original href didn't have a trailing slash
# and the base_url didn't end with a slash
base_parsed = urlparse(base_url)
if not href.strip().endswith('/') and not base_parsed.path.endswith('/'):
path = path.rstrip('/')
# Add trailing slash for URLs without explicit paths (indicates directory)
# But skip this for special protocols that don't use standard URL structure
elif not path:
special_protocols = {"javascript:", "mailto:", "tel:", "file:", "data:"}
if not any(href.strip().lower().startswith(p) for p in special_protocols):
path = '/'
# ── query ──
query = parsed.query
@@ -2221,6 +2249,8 @@ def normalize_url(
}
if extra_drop_params:
default_tracking |= {p.lower() for p in extra_drop_params}
if params_to_remove:
default_tracking |= {p.lower() for p in params_to_remove}
params = [(k, v) for k, v in params if k not in default_tracking]
if sort_query:
@@ -2229,7 +2259,10 @@ def normalize_url(
query = urlencode(params, doseq=True) if params else ''
# ── fragment ──
fragment = parsed.fragment if keep_fragment else ''
if remove_fragments is True:
fragment = ''
else:
fragment = parsed.fragment if keep_fragment else ''
# Re-assemble
normalized = urlunparse((
@@ -2453,9 +2486,19 @@ def is_external_url(url: str, base_domain: str) -> bool:
if not parsed.netloc: # Relative URL
return False
# Strip 'www.' from both domains for comparison
url_domain = parsed.netloc.lower().replace("www.", "")
base = base_domain.lower().replace("www.", "")
# Don't strip 'www.' from domains for comparison - treat www.example.com and example.com as different
url_domain = parsed.netloc.lower()
base = base_domain.lower()
# Strip user credentials from URL domain
if '@' in url_domain:
url_domain = url_domain.split('@', 1)[1]
# Strip ports from both for comparison (any port should be considered same domain)
if ':' in url_domain:
url_domain = url_domain.rsplit(':', 1)[0]
if ':' in base:
base = base.rsplit(':', 1)[0]
# Check if URL domain ends with base domain
return not url_domain.endswith(base)