diff --git a/crawl4ai/async_webcrawler.py b/crawl4ai/async_webcrawler.py index c722235b..7b83d2d5 100644 --- a/crawl4ai/async_webcrawler.py +++ b/crawl4ai/async_webcrawler.py @@ -16,7 +16,6 @@ from .models import ( MarkdownGenerationResult, CrawlerTaskResult, DispatchResult, - DeepCrawlingProgress, ) from .async_database import async_db_manager from .chunking_strategy import * # noqa: F403 diff --git a/docs/scraper/deep_crawl_quickstart.py b/docs/scraper/deep_crawl_quickstart.py index 1e97714e..63a0bf9b 100644 --- a/docs/scraper/deep_crawl_quickstart.py +++ b/docs/scraper/deep_crawl_quickstart.py @@ -135,6 +135,7 @@ async def advanced_scraper_example(): try: # Use streaming mode + results = [] result_generator = await crawler.adeep_crawl( "https://techcrunch.com", strategy=bfs_strategy, @@ -151,6 +152,7 @@ async def advanced_scraper_example(): logger.info( f"Processed at depth: {result.depth} with score: {result.score:.3f} : \n {result.url}" ) + results.append(result) else: stats["errors"] += 1 logger.error(