Skip to content

Commit 0d2b7b3

Browse files
committed
update dependenceis
1 parent 437e48f commit 0d2b7b3

File tree

3 files changed

+22
-23
lines changed

3 files changed

+22
-23
lines changed

pyproject.toml

Lines changed: 12 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,7 @@
11
[project]
22
name = "scrapegraphai"
3-
4-
53
version = "1.14.1b1"
6-
7-
84
description = "A web scraping library based on LangChain which uses LLM and direct graph logic to create scraping pipelines."
9-
105
authors = [
116
{ name = "Marco Vinciguerra", email = "mvincig11@gmail.com" },
127
{ name = "Marco Perini", email = "perinim.98@gmail.com" },
@@ -15,32 +10,23 @@ authors = [
1510

1611
dependencies = [
1712
"langchain>=0.2.14",
18-
"langchain-fireworks>=0.1.3",
19-
"langchain_community>=0.2.9",
2013
"langchain-google-genai>=1.0.7",
21-
"langchain-google-vertexai>=1.0.7",
2214
"langchain-openai>=0.1.22",
23-
"langchain-groq>=0.1.3",
24-
"langchain-aws>=0.1.3",
25-
"langchain-anthropic>=0.1.11",
2615
"langchain-mistralai>=0.1.12",
27-
"langchain-huggingface>=0.0.3",
28-
"langchain-nvidia-ai-endpoints>=0.1.6",
16+
"langchain_community>=0.2.9",
17+
"langchain-aws>=0.1.3",
2918
"html2text>=2024.2.26",
3019
"faiss-cpu>=1.8.0",
3120
"beautifulsoup4>=4.12.3",
3221
"pandas>=2.2.2",
3322
"python-dotenv>=1.0.1",
3423
"tiktoken>=0.7",
3524
"tqdm>=4.66.4",
36-
"graphviz>=0.20.3",
3725
"minify-html>=0.15.0",
3826
"free-proxy>=1.1.1",
3927
"playwright>=1.43.0",
4028
"google>=3.0.0",
4129
"undetected-playwright>=0.3.0",
42-
"semchunk>=1.0.1",
43-
"browserbase>=0.3.0",
4430
]
4531

4632
license = "MIT"
@@ -78,6 +64,16 @@ requires-python = ">=3.9,<4.0"
7864
[project.optional-dependencies]
7965
burr = ["burr[start]==0.22.1"]
8066
docs = ["sphinx==6.0", "furo==2024.5.6"]
67+
other = [
68+
"graphviz>=0.20.3",
69+
"langchain-fireworks>=0.1.3",
70+
"langchain-groq>=0.1.3",
71+
"langchain-anthropic>=0.1.11",
72+
"langchain-huggingface>=0.0.3",
73+
"langchain-nvidia-ai-endpoints>=0.1.6",
74+
"semchunk>=1.0.1",
75+
"browserbase>=0.3.0",
76+
]
8177

8278
[build-system]
8379
requires = ["hatchling"]

scrapegraphai/graphs/abstract_graph.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@
77
import uuid
88
import warnings
99
from pydantic import BaseModel
10-
from langchain_community.chat_models import ErnieBotChat
11-
from langchain_nvidia_ai_endpoints import ChatNVIDIA
1210
from langchain.chat_models import init_chat_model
1311
from ..helpers import models_tokens
1412
from ..models import (
@@ -147,16 +145,17 @@ def handle_model(model_name, provider, token_key, default_token=8192):
147145
warnings.simplefilter("ignore")
148146
return init_chat_model(**llm_params)
149147

150-
known_models = ["chatgpt","gpt","openai", "azure_openai", "google_genai", "ollama", "oneapi", "nvidia", "groq", "google_vertexai", "bedrock", "mistralai", "hugging_face", "deepseek", "ernie", "fireworks"]
148+
known_models = ["chatgpt","gpt","openai", "azure_openai", "google_genai",
149+
"ollama", "oneapi", "nvidia", "groq", "google_vertexai",
150+
"bedrock", "mistralai", "hugging_face", "deepseek", "ernie", "fireworks"]
151151

152152
if llm_params["model"].split("/")[0] not in known_models and llm_params["model"].split("-")[0] not in known_models:
153153
raise ValueError(f"Model '{llm_params['model']}' is not supported")
154154

155155
try:
156156
if "azure" in llm_params["model"]:
157157
model_name = llm_params["model"].split("/")[-1]
158-
return handle_model(model_name, "azure_openai", model_name)
159-
158+
return handle_model(model_name, "azure_openai", model_name)
160159
if "fireworks" in llm_params["model"]:
161160
model_name = "/".join(llm_params["model"].split("/")[1:])
162161
token_key = llm_params["model"].split("/")[-1]
@@ -188,7 +187,6 @@ def handle_model(model_name, provider, token_key, default_token=8192):
188187
model_name = llm_params["model"].split("/")[-1]
189188
return handle_model(model_name, "mistralai", model_name)
190189

191-
# Instantiate the language model based on the model name (models that do not use the common interface)
192190
elif "deepseek" in llm_params["model"]:
193191
try:
194192
self.model_token = models_tokens["deepseek"][llm_params["model"]]
@@ -198,6 +196,8 @@ def handle_model(model_name, provider, token_key, default_token=8192):
198196
return DeepSeek(llm_params)
199197

200198
elif "ernie" in llm_params["model"]:
199+
from langchain_community.chat_models import ErnieBotChat
200+
201201
try:
202202
self.model_token = models_tokens["ernie"][llm_params["model"]]
203203
except KeyError:
@@ -215,6 +215,8 @@ def handle_model(model_name, provider, token_key, default_token=8192):
215215
return OneApi(llm_params)
216216

217217
elif "nvidia" in llm_params["model"]:
218+
from langchain_nvidia_ai_endpoints import ChatNVIDIA
219+
218220
try:
219221
self.model_token = models_tokens["nvidia"][llm_params["model"].split("/")[-1]]
220222
llm_params["model"] = "/".join(llm_params["model"].split("/")[1:])

scrapegraphai/nodes/fetch_node.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,6 @@
1010
from langchain_core.documents import Document
1111
from ..utils.cleanup_html import cleanup_html
1212
from ..docloaders import ChromiumLoader
13-
from ..docloaders.browser_base import browser_base_fetch
1413
from ..utils.convert_to_md import convert_to_md
1514
from ..utils.logging import get_logger
1615
from .base_node import BaseNode
@@ -269,6 +268,8 @@ def handle_web_source(self, state, source):
269268
loader_kwargs = self.node_config.get("loader_kwargs", {})
270269

271270
if self.browser_base is not None:
271+
from ..docloaders.browser_base import browser_base_fetch
272+
272273
data = browser_base_fetch(self.browser_base.get("api_key"),
273274
self.browser_base.get("project_id"), [source])
274275

0 commit comments

Comments
 (0)