Skip to content

Commit baa6006

Browse files
fix(graphgen): fix lint errors
1 parent eb19b8d commit baa6006

File tree

4 files changed

+36
-29
lines changed

4 files changed

+36
-29
lines changed

graphgen/operators/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,4 +10,4 @@
1010
"judge_statement",
1111
"search_wikipedia",
1212
"traverse_graph_by_edge"
13-
]
13+
]

graphgen/operators/extract_kg.py

Lines changed: 11 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -3,14 +3,14 @@
33

44
from typing import List
55
from collections import defaultdict
6+
from tqdm.asyncio import tqdm as tqdm_async
67
from models import Chunk, OpenAIModel, Tokenizer
78
from models.storage.base_storage import BaseGraphStorage
89
from templates import KG_EXTRACTION_PROMPT
9-
from tqdm.asyncio import tqdm as tqdm_async
1010
from utils import (logger, pack_history_conversations, split_string_by_multi_markers,
1111
handle_single_entity_extraction, handle_single_relationship_extraction,
1212
detect_if_chinese)
13-
from .merge_kg import merge_nodes, merge_edges
13+
from graphgen.operators.merge_kg import merge_nodes, merge_edges
1414

1515

1616
async def extract_kg(
@@ -51,12 +51,18 @@ async def _process_single_content(chunk: Chunk, max_loop: int = 3):
5151

5252
history = pack_history_conversations(hint_prompt, final_result)
5353
for loop_index in range(max_loop):
54-
if_loop_result = await llm_client.generate_answer(text=KG_EXTRACTION_PROMPT[language]["IF_LOOP"], history=history)
54+
if_loop_result = await llm_client.generate_answer(
55+
text=KG_EXTRACTION_PROMPT[language]["IF_LOOP"],
56+
history=history
57+
)
5558
if_loop_result = if_loop_result.strip().strip('"').strip("'").lower()
5659
if if_loop_result != "yes":
5760
break
5861

59-
glean_result = await llm_client.generate_answer(text=KG_EXTRACTION_PROMPT[language]["CONTINUE"], history=history)
62+
glean_result = await llm_client.generate_answer(
63+
text=KG_EXTRACTION_PROMPT[language]["CONTINUE"],
64+
history=history
65+
)
6066
logger.info(f"Loop {loop_index} glean: {glean_result}")
6167

6268
history += pack_history_conversations(KG_EXTRACTION_PROMPT[language]["CONTINUE"], glean_result)
@@ -101,7 +107,7 @@ async def _process_single_content(chunk: Chunk, max_loop: int = 3):
101107
):
102108
try:
103109
results.append(await result)
104-
except Exception as e:
110+
except Exception as e: # pylint: disable=broad-except
105111
logger.error("Error occurred while extracting entities and relationships from chunks: %s", e)
106112

107113
nodes = defaultdict(list)

graphgen/operators/merge_kg.py

Lines changed: 21 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,12 @@
11
from collections import Counter
2+
import asyncio
3+
from tqdm.asyncio import tqdm as tqdm_async
4+
25
from utils.format import split_string_by_multi_markers
6+
from utils import logger, detect_main_language
37
from models import TopkTokenModel, Tokenizer
48
from models.storage.base_storage import BaseGraphStorage
59
from templates import KG_SUMMARIZATION_PROMPT, KG_EXTRACTION_PROMPT
6-
from utils import logger, detect_main_language
7-
from tqdm.asyncio import tqdm as tqdm_async
8-
import asyncio
910

1011
async def _handle_kg_summary(
1112
entity_or_relation_name: str,
@@ -100,11 +101,11 @@ async def process_single_node(entity_name: str, node_data: list[dict]):
100101
set([dp["source_id"] for dp in node_data] + source_ids)
101102
)
102103

103-
node_data = dict(
104-
entity_type=entity_type,
105-
description=description,
106-
source_id=source_id
107-
)
104+
node_data = {
105+
"entity_type": entity_type,
106+
"description": description,
107+
"source_id": source_id
108+
}
108109
await kg_instance.upsert_node(
109110
entity_name,
110111
node_data=node_data
@@ -124,7 +125,7 @@ async def process_single_node(entity_name: str, node_data: list[dict]):
124125
):
125126
try:
126127
entities_data.append(await result)
127-
except Exception as e:
128+
except Exception as e: # pylint: disable=broad-except
128129
logger.error("Error occurred while inserting entities into storage: %s", e)
129130

130131

@@ -168,7 +169,7 @@ async def process_single_edge(src_id: str, tgt_id: str, edge_data: list[dict]):
168169
)
169170

170171
for insert_id in [src_id, tgt_id]:
171-
if not (await kg_instance.has_node(insert_id)):
172+
if not await kg_instance.has_node(insert_id):
172173
await kg_instance.upsert_node(
173174
insert_id,
174175
node_data={
@@ -185,17 +186,17 @@ async def process_single_edge(src_id: str, tgt_id: str, edge_data: list[dict]):
185186
await kg_instance.upsert_edge(
186187
src_id,
187188
tgt_id,
188-
edge_data=dict(
189-
description=description,
190-
source_id=source_id
191-
)
189+
edge_data={
190+
"source_id": source_id,
191+
"description": description
192+
}
192193
)
193194

194-
edge_data = dict(
195-
src_id=src_id,
196-
tgt_id=tgt_id,
197-
description=description,
198-
)
195+
edge_data = {
196+
"src_id": src_id,
197+
"tgt_id": tgt_id,
198+
"description": description
199+
}
199200
return edge_data
200201

201202
logger.info("Inserting relationships into storage...")
@@ -210,5 +211,5 @@ async def process_single_edge(src_id: str, tgt_id: str, edge_data: list[dict]):
210211
):
211212
try:
212213
relationships_data.append(await result)
213-
except Exception as e:
214+
except Exception as e: # pylint: disable=broad-except
214215
logger.error("Error occurred while inserting relationships into storage: %s", e)

models/storage/networkx_storage.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ async def update_node(self, node_id: str, node_data: dict[str, str]):
126126
if self._graph.has_node(node_id):
127127
self._graph.nodes[node_id].update(node_data)
128128
else:
129-
logger.warning(f"Node {node_id} not found in the graph for update.")
129+
logger.warning("Node %s not found in the graph for update.", node_id)
130130

131131
async def upsert_edge(
132132
self, source_node_id: str, target_node_id: str, edge_data: dict[str, str]
@@ -147,6 +147,6 @@ async def delete_node(self, node_id: str):
147147
"""
148148
if self._graph.has_node(node_id):
149149
self._graph.remove_node(node_id)
150-
logger.info(f"Node {node_id} deleted from the graph.")
150+
logger.info("Node %s deleted from the graph.", node_id)
151151
else:
152-
logger.warning(f"Node {node_id} not found in the graph for deletion.")
152+
logger.warning("Node %s not found in the graph for deletion.", node_id)

0 commit comments

Comments
 (0)