Skip to content

Commit

Permalink
fix: fetch node
Browse files Browse the repository at this point in the history
  • Loading branch information
VinciGit00 committed Aug 10, 2024
1 parent caff63b commit f01b55e
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 75 deletions.
36 changes: 0 additions & 36 deletions requirements-dev.lock
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,6 @@
# features: []
# all-features: false
# with-sources: false
# generate-hashes: false
# universal: false

-e file:.
aiofiles==24.1.0
Expand Down Expand Up @@ -112,7 +110,6 @@ filelock==3.15.4
# via huggingface-hub
# via torch
# via transformers
# via triton
fireworks-ai==0.14.0
# via langchain-fireworks
fonttools==4.53.1
Expand Down Expand Up @@ -188,7 +185,6 @@ graphviz==0.20.3
# via scrapegraphai
greenlet==3.0.3
# via playwright
# via sqlalchemy
groq==0.9.0
# via langchain-groq
grpc-google-iam-v1==0.13.1
Expand Down Expand Up @@ -362,34 +358,6 @@ numpy==1.26.4
# via shapely
# via streamlit
# via transformers
nvidia-cublas-cu12==12.1.3.1
# via nvidia-cudnn-cu12
# via nvidia-cusolver-cu12
# via torch
nvidia-cuda-cupti-cu12==12.1.105
# via torch
nvidia-cuda-nvrtc-cu12==12.1.105
# via torch
nvidia-cuda-runtime-cu12==12.1.105
# via torch
nvidia-cudnn-cu12==8.9.2.26
# via torch
nvidia-cufft-cu12==11.0.2.54
# via torch
nvidia-curand-cu12==10.3.2.106
# via torch
nvidia-cusolver-cu12==11.4.5.107
# via torch
nvidia-cusparse-cu12==12.1.0.106
# via nvidia-cusolver-cu12
# via torch
nvidia-nccl-cu12==2.19.3
# via torch
nvidia-nvjitlink-cu12==12.6.20
# via nvidia-cusolver-cu12
# via nvidia-cusparse-cu12
nvidia-nvtx-cu12==12.1.105
# via torch
openai==1.37.0
# via burr
# via langchain-fireworks
Expand Down Expand Up @@ -631,8 +599,6 @@ tqdm==4.66.4
transformers==4.43.3
# via langchain-huggingface
# via sentence-transformers
triton==2.2.0
# via torch
typer==0.12.3
# via fastapi-cli
typing-extensions==4.12.2
Expand Down Expand Up @@ -676,8 +642,6 @@ uvicorn==0.30.3
# via fastapi
uvloop==0.19.0
# via uvicorn
watchdog==4.0.1
# via streamlit
watchfiles==0.22.0
# via uvicorn
websockets==12.0
Expand Down
34 changes: 0 additions & 34 deletions requirements.lock
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,6 @@
# features: []
# all-features: false
# with-sources: false
# generate-hashes: false
# universal: false

-e file:.
aiohttp==3.9.5
Expand Down Expand Up @@ -69,7 +67,6 @@ filelock==3.15.4
# via huggingface-hub
# via torch
# via transformers
# via triton
fireworks-ai==0.14.0
# via langchain-fireworks
free-proxy==1.1.1
Expand Down Expand Up @@ -136,7 +133,6 @@ graphviz==0.20.3
# via scrapegraphai
greenlet==3.0.3
# via playwright
# via sqlalchemy
groq==0.9.0
# via langchain-groq
grpc-google-iam-v1==0.13.1
Expand Down Expand Up @@ -267,34 +263,6 @@ numpy==1.26.4
# via sentence-transformers
# via shapely
# via transformers
nvidia-cublas-cu12==12.1.3.1
# via nvidia-cudnn-cu12
# via nvidia-cusolver-cu12
# via torch
nvidia-cuda-cupti-cu12==12.1.105
# via torch
nvidia-cuda-nvrtc-cu12==12.1.105
# via torch
nvidia-cuda-runtime-cu12==12.1.105
# via torch
nvidia-cudnn-cu12==8.9.2.26
# via torch
nvidia-cufft-cu12==11.0.2.54
# via torch
nvidia-curand-cu12==10.3.2.106
# via torch
nvidia-cusolver-cu12==11.4.5.107
# via torch
nvidia-cusparse-cu12==12.1.0.106
# via nvidia-cusolver-cu12
# via torch
nvidia-nccl-cu12==2.19.3
# via torch
nvidia-nvjitlink-cu12==12.6.20
# via nvidia-cusolver-cu12
# via nvidia-cusparse-cu12
nvidia-nvtx-cu12==12.1.105
# via torch
openai==1.37.0
# via langchain-fireworks
# via langchain-openai
Expand Down Expand Up @@ -446,8 +414,6 @@ tqdm==4.66.4
transformers==4.43.3
# via langchain-huggingface
# via sentence-transformers
triton==2.2.0
# via torch
typing-extensions==4.12.2
# via anthropic
# via anyio
Expand Down
10 changes: 5 additions & 5 deletions scrapegraphai/nodes/fetch_node.py
Original file line number Diff line number Diff line change
Expand Up @@ -260,7 +260,7 @@ def handle_web_source(self, state, source):

if (isinstance(self.llm_model, ChatOpenAI)
and not self.script_creator) or (self.force and not self.script_creator):
parsed_content = convert_to_md(source, input_data[0])
parsed_content = convert_to_md(source, parsed_content)

compressed_document = [Document(page_content=parsed_content)]
else:
Expand Down Expand Up @@ -288,14 +288,14 @@ def handle_web_source(self, state, source):
parsed_content = document[0].page_content

if isinstance(self.llm_model, ChatOpenAI) and not self.script_creator or self.force and not self.script_creator and not self.openai_md_enabled:
parsed_content = convert_to_md(document[0].page_content, input_data[0])
parsed_content = convert_to_md(document[0].page_content, parsed_content)

compressed_document = [
Document(page_content=parsed_content, metadata={"source": "html file"})
]

return self.update_state(state, compressed_document)

def update_state(self, state, compressed_document):
"""
Updates the state with the output data from the node.
Expand All @@ -308,6 +308,6 @@ def update_state(self, state, compressed_document):
Returns:
dict: The updated state with the output data.
"""

state.update({self.output[0]: compressed_document,})
return state
return state

0 comments on commit f01b55e

Please sign in to comment.