Skip to content

Commit 2eaeba3

Browse files
committed
refactor: load() logging should be info instead of debug
1 parent fbc4b20 commit 2eaeba3

File tree

2 files changed

+16
-16
lines changed

2 files changed

+16
-16
lines changed

CHANGELOG.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,10 @@
11
## [1.1.3](https://github.com/lpm0073/hybrid-search-retriever/compare/v1.1.2...v1.1.3) (2023-12-02)
22

3-
43
### Bug Fixes
54

6-
* add langchain-experimental for yt example ([f9d6d6d](https://github.com/lpm0073/hybrid-search-retriever/commit/f9d6d6d0b11ff9c1f06faf7eb69511bc5702066d))
7-
* correct type error with DEBUG_MODE ([a96bdfd](https://github.com/lpm0073/hybrid-search-retriever/commit/a96bdfdb5a0b015740110e02f9f9b06917cd31c7))
8-
* move retriever results to system_message ([203c8b3](https://github.com/lpm0073/hybrid-search-retriever/commit/203c8b300cda156ac44a0c6e02510c2ab6a2b074))
5+
- add langchain-experimental for yt example ([f9d6d6d](https://github.com/lpm0073/hybrid-search-retriever/commit/f9d6d6d0b11ff9c1f06faf7eb69511bc5702066d))
6+
- correct type error with DEBUG_MODE ([a96bdfd](https://github.com/lpm0073/hybrid-search-retriever/commit/a96bdfdb5a0b015740110e02f9f9b06917cd31c7))
7+
- move retriever results to system_message ([203c8b3](https://github.com/lpm0073/hybrid-search-retriever/commit/203c8b300cda156ac44a0c6e02510c2ab6a2b074))
98

109
## [1.1.2](https://github.com/lpm0073/hybrid-search-retriever/compare/v1.1.1...v1.1.2) (2023-12-01)
1110

models/hybrid_search_retreiver.py

Lines changed: 13 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ def create_documents(self, texts):
7373

7474

7575
class HybridSearchRetriever:
76-
"""Hybrid Search Retriever (OpenAI + Pinecone)"""
76+
"""Hybrid Search Retriever"""
7777

7878
_chat: ChatOpenAI = None
7979
_openai_embeddings: OpenAIEmbeddings = None
@@ -182,16 +182,16 @@ def load(self, filepath: str):
182182
https://docs.pinecone.io/docs/manage-indexes#selective-metadata-indexing
183183
"""
184184
try:
185-
logging.debug("Deleting index...")
185+
logging.info("Deleting index...")
186186
pinecone.delete_index(Config.PINECONE_INDEX_NAME)
187187
except pinecone.exceptions.PineconeException:
188-
logging.debug("Index does not exist. Continuing...")
188+
logging.info("Index does not exist. Continuing...")
189189

190190
metadata_config = {
191191
"indexed": [Config.PINECONE_VECTORSTORE_TEXT_KEY, "lc_type"],
192192
"context": ["lc_text"],
193193
}
194-
logging.debug("Creating index. This may take a few minutes...")
194+
logging.info("Creating index. This may take a few minutes...")
195195
pinecone.create_index(
196196
Config.PINECONE_INDEX_NAME,
197197
dimension=Config.PINECONE_DIMENSIONS,
@@ -204,23 +204,23 @@ def load(self, filepath: str):
204204
for pdf_file in pdf_files:
205205
i += 1
206206
j = len(pdf_files)
207-
logging.debug("Loading PDF %s of %s: %s", i, j, pdf_file)
207+
logging.info("Loading PDF %s of %s: %s", i, j, pdf_file)
208208
loader = PyPDFLoader(file_path=pdf_file)
209209
docs = loader.load()
210210
k = 0
211211
for doc in docs:
212212
k += 1
213-
logging.debug(k * "-", end="\r")
213+
logging.info(k * "-", end="\r")
214214
documents = self.text_splitter.create_documents([doc.page_content])
215215
document_texts = [doc.page_content for doc in documents]
216216
embeddings = self.openai_embeddings.embed_documents(document_texts)
217217
self.vector_store.add_documents(documents=documents, embeddings=embeddings)
218218

219-
logging.debug("Finished loading PDFs")
219+
logging.info("Finished loading PDFs")
220220

221221
def rag(self, human_message: Union[str, HumanMessage]):
222222
"""
223-
Embedded prompt.
223+
Retrieval Augmented Generation prompt.
224224
1. Retrieve human message prompt: Given a user input, relevant splits are retrieved
225225
from storage using a Retriever.
226226
2. Generate: A ChatModel / LLM produces an answer using a prompt that includes
@@ -265,9 +265,10 @@ def rag(self, human_message: Union[str, HumanMessage]):
265265
response = self.cached_chat_request(system_message=system_message, human_message=human_message)
266266

267267
logging.debug("------------------------------------------------------")
268-
logging.debug("Retrieved %i related documents from Pinecone", len(documents))
269-
logging.debug("System messages contains %i words", len(system_message.content.split()))
270-
logging.debug("Prompt: %s", system_message.content)
271-
logging.debug("Response:")
268+
logging.debug("rag() Retrieval Augmented Generation prompt")
269+
logging.debug("Diagnostic information:")
270+
logging.debug(" Retrieved %i related documents from Pinecone", len(documents))
271+
logging.debug(" System messages contains %i words", len(system_message.content.split()))
272+
logging.debug(" Prompt: %s", system_message.content)
272273
logging.debug("------------------------------------------------------")
273274
return response.content

0 commit comments

Comments
 (0)