## [0.2.71] 2024-06-26
• Refactored `crawler_strategy.py` to handle exceptions and improve error messages • Improved `get_content_of_website_optimized` function in `utils.py` for better performance • Updated `utils.py` with latest changes • Migrated to `ChromeDriverManager` for resolving Chrome driver download issues
This commit is contained in:
4
.gitignore
vendored
4
.gitignore
vendored
@@ -185,4 +185,6 @@ local/
|
|||||||
|
|
||||||
a.txt
|
a.txt
|
||||||
.lambda_function.py
|
.lambda_function.py
|
||||||
ec2*
|
ec2*
|
||||||
|
|
||||||
|
update_changelog.sh
|
||||||
45
CHANGELOG.md
45
CHANGELOG.md
@@ -1,5 +1,16 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [0.2.71] 2024-06-26
|
||||||
|
• Refactored `crawler_strategy.py` to handle exceptions and improve error messages
|
||||||
|
• Improved `get_content_of_website_optimized` function in `utils.py` for better performance
|
||||||
|
• Updated `utils.py` with latest changes
|
||||||
|
• Migrated to `ChromeDriverManager` for resolving Chrome driver download issues
|
||||||
|
|
||||||
|
## [0.2.71] - 2024-06-25
|
||||||
|
### Fixed
|
||||||
|
- Speed up twice the extraction function.
|
||||||
|
|
||||||
|
|
||||||
## [0.2.6] - 2024-06-22
|
## [0.2.6] - 2024-06-22
|
||||||
### Fixed
|
### Fixed
|
||||||
- Fix issue #19: Update Dockerfile to ensure compatibility across multiple platforms.
|
- Fix issue #19: Update Dockerfile to ensure compatibility across multiple platforms.
|
||||||
@@ -20,37 +31,3 @@
|
|||||||
## [0.2.4] - 2024-06-17
|
## [0.2.4] - 2024-06-17
|
||||||
### Fixed
|
### Fixed
|
||||||
- Fix issue #22: Use MD5 hash for caching HTML files to handle long URLs
|
- Fix issue #22: Use MD5 hash for caching HTML files to handle long URLs
|
||||||
## Update 2024-06-26
|
|
||||||
|
|
||||||
### Commits in the last 3 hours:
|
|
||||||
4756d0a - Refactor crawler_strategy.py to handle exceptions and improve error messages
|
|
||||||
7ba2142 - chore: Refactor get_content_of_website_optimized function in utils.py
|
|
||||||
96d1eb0 - Some updated ins utils.py
|
|
||||||
144cfa0 - Switch to ChromeDriverManager due some issues with download the chrome driver
|
|
||||||
null
|
|
||||||
null
|
|
||||||
null
|
|
||||||
Here is a rewritten version of the changelog update in a nicer and more condensed way:
|
|
||||||
|
|
||||||
**Update 2024-06-26**
|
|
||||||
|
|
||||||
We've made some exciting improvements to our codebase! Here are the highlights:
|
|
||||||
|
|
||||||
* Refactored our crawler strategy to handle exceptions and provide clearer error messages
|
|
||||||
* Optimized our content retrieval function for improved performance
|
|
||||||
* Updated internal utilities for better functionality
|
|
||||||
* Switched to ChromeDriverManager to resolve issues with downloading Chrome drivers
|
|
||||||
|
|
||||||
These updates aim to improve stability, reliability, and overall performance. Thank you for using our tool!
|
|
||||||
Here is a rewritten version of the changelog update:
|
|
||||||
|
|
||||||
**June 26, 2024**
|
|
||||||
|
|
||||||
We've made some improvements to our code to make it more reliable and user-friendly!
|
|
||||||
|
|
||||||
In the last 3 hours, we've committed 4 changes:
|
|
||||||
|
|
||||||
* Improved error handling and messaging in [crawler_strategy.py](https://example.com/crawler_strategy.py)
|
|
||||||
* Refactored [get_content_of_website_optimized](https://example.com/utils.py) in [utils.py](https://example.com/utils.py)
|
|
||||||
* Made updates to [utils.py](https://example.com/utils.py)
|
|
||||||
* Switched to [ChromeDriverManager](https://example.com/ChromeDriverManager) to resolve issues with downloading the Chrome driver.
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# Crawl4AI v0.2.7 🕷️🤖
|
# Crawl4AI v0.2.71 🕷️🤖
|
||||||
|
|
||||||
[](https://github.com/unclecode/crawl4ai/stargazers)
|
[](https://github.com/unclecode/crawl4ai/stargazers)
|
||||||
[](https://github.com/unclecode/crawl4ai/network/members)
|
[](https://github.com/unclecode/crawl4ai/network/members)
|
||||||
|
|||||||
@@ -1,6 +1,12 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
## [0.2.7] - 2024-06-27
|
## [0.2.71] 2024-06-26
|
||||||
|
• Refactored `crawler_strategy.py` to handle exceptions and improve error messages
|
||||||
|
• Improved `get_content_of_website_optimized` function in `utils.py` for better performance
|
||||||
|
• Updated `utils.py` with latest changes
|
||||||
|
• Migrated to `ChromeDriverManager` for resolving Chrome driver download issues
|
||||||
|
|
||||||
|
## [0.2.71] - 2024-06-25
|
||||||
### Fixed
|
### Fixed
|
||||||
- Speed up twice the extraction function.
|
- Speed up twice the extraction function.
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# Crawl4AI v0.2.7
|
# Crawl4AI v0.2.71
|
||||||
|
|
||||||
Welcome to the official documentation for Crawl4AI! 🕷️🤖 Crawl4AI is an open-source Python library designed to simplify web crawling and extract useful information from web pages. This documentation will guide you through the features, usage, and customization of Crawl4AI.
|
Welcome to the official documentation for Crawl4AI! 🕷️🤖 Crawl4AI is an open-source Python library designed to simplify web crawling and extract useful information from web pages. This documentation will guide you through the features, usage, and customization of Crawl4AI.
|
||||||
|
|
||||||
|
|||||||
2
setup.py
2
setup.py
@@ -33,7 +33,7 @@ class CustomInstallCommand(install):
|
|||||||
|
|
||||||
setup(
|
setup(
|
||||||
name="Crawl4AI",
|
name="Crawl4AI",
|
||||||
version="0.2.7",
|
version="0.2.71",
|
||||||
description="🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper",
|
description="🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper",
|
||||||
long_description=open("README.md").read(),
|
long_description=open("README.md").read(),
|
||||||
long_description_content_type="text/markdown",
|
long_description_content_type="text/markdown",
|
||||||
|
|||||||
Reference in New Issue
Block a user