chore: Update torch dependency to version 2.3.0
This commit is contained in:
@@ -16,4 +16,4 @@ selenium==4.20.0
|
|||||||
uvicorn==0.29.0
|
uvicorn==0.29.0
|
||||||
transformers==4.40.2
|
transformers==4.40.2
|
||||||
chromedriver-autoinstaller==0.6.4
|
chromedriver-autoinstaller==0.6.4
|
||||||
torch
|
torch==2.3.0
|
||||||
59
setup.py
59
setup.py
@@ -3,7 +3,10 @@ from setuptools import setup, find_packages
|
|||||||
# Read the requirements from requirements.txt
|
# Read the requirements from requirements.txt
|
||||||
with open("requirements.txt") as f:
|
with open("requirements.txt") as f:
|
||||||
requirements = f.read().splitlines()
|
requirements = f.read().splitlines()
|
||||||
|
|
||||||
|
# Define the requirements for different environments
|
||||||
|
requirements_without_torch = [req for req in requirements if not req.startswith("torch")]
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
name="Crawl4AI",
|
name="Crawl4AI",
|
||||||
version="0.1.2",
|
version="0.1.2",
|
||||||
@@ -16,6 +19,9 @@ setup(
|
|||||||
license="MIT",
|
license="MIT",
|
||||||
packages=find_packages(),
|
packages=find_packages(),
|
||||||
install_requires=requirements,
|
install_requires=requirements,
|
||||||
|
extras_require={
|
||||||
|
"colab": requirements_without_torch, # Exclude torch for Colab
|
||||||
|
},
|
||||||
entry_points={
|
entry_points={
|
||||||
'console_scripts': [
|
'console_scripts': [
|
||||||
'crawl4ai-download-models=crawl4ai.model_loader:main',
|
'crawl4ai-download-models=crawl4ai.model_loader:main',
|
||||||
@@ -33,54 +39,3 @@ setup(
|
|||||||
],
|
],
|
||||||
python_requires=">=3.7",
|
python_requires=">=3.7",
|
||||||
)
|
)
|
||||||
|
|
||||||
from setuptools import setup, find_packages
|
|
||||||
from setuptools.command.install import install as _install
|
|
||||||
import subprocess
|
|
||||||
import sys
|
|
||||||
|
|
||||||
class InstallCommand(_install):
|
|
||||||
def run(self):
|
|
||||||
# Run the standard install first
|
|
||||||
_install.run(self)
|
|
||||||
# Now handle the dependencies manually
|
|
||||||
self.manual_dependencies_install()
|
|
||||||
|
|
||||||
def manual_dependencies_install(self):
|
|
||||||
with open('requirements.txt') as f:
|
|
||||||
dependencies = f.read().splitlines()
|
|
||||||
for dependency in dependencies:
|
|
||||||
subprocess.check_call([sys.executable, '-m', 'pip', 'install', dependency])
|
|
||||||
|
|
||||||
# setup(
|
|
||||||
# name="Crawl4AI",
|
|
||||||
# version="0.1.0",
|
|
||||||
# description="🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper",
|
|
||||||
# long_description=open("README.md").read(),
|
|
||||||
# long_description_content_type="text/markdown",
|
|
||||||
# url="https://github.com/unclecode/crawl4ai",
|
|
||||||
# author="Unclecode",
|
|
||||||
# author_email="unclecode@kidocode.com",
|
|
||||||
# license="MIT",
|
|
||||||
# packages=find_packages(),
|
|
||||||
# install_requires=[], # Leave this empty to avoid default dependency resolution
|
|
||||||
# cmdclass={
|
|
||||||
# 'install': InstallCommand,
|
|
||||||
# },
|
|
||||||
# entry_points={
|
|
||||||
# 'console_scripts': [
|
|
||||||
# 'crawl4ai-download-models=crawl4ai.model_loader:main',
|
|
||||||
# ],
|
|
||||||
# },
|
|
||||||
# classifiers=[
|
|
||||||
# "Development Status :: 3 - Alpha",
|
|
||||||
# "Intended Audience :: Developers",
|
|
||||||
# "License :: OSI Approved :: Apache Software License",
|
|
||||||
# "Programming Language :: Python :: 3",
|
|
||||||
# "Programming Language :: Python :: 3.7",
|
|
||||||
# "Programming Language :: Python :: 3.8",
|
|
||||||
# "Programming Language :: Python :: 3.9",
|
|
||||||
# "Programming Language :: Python :: 3.10",
|
|
||||||
# ],
|
|
||||||
# python_requires=">=3.7",
|
|
||||||
# )
|
|
||||||
|
|||||||
Reference in New Issue
Block a user