File tree Expand file tree Collapse file tree 3 files changed +9
-8
lines changed Expand file tree Collapse file tree 3 files changed +9
-8
lines changed Original file line number Diff line number Diff line change @@ -126,12 +126,12 @@ def upload_and_process_pdf() -> list:
126
126
return docs
127
127
128
128
129
- @lru_cache (maxsize = 20 )
130
129
def get_response (query : str ) -> str :
130
+ app_logger .info (f'\033 [36mUser Query: { query } \033 [0m' )
131
131
try :
132
132
if model is not None :
133
133
response = model .run (query )
134
- app_logger .info (f'llm response : { response } ' )
134
+ app_logger .info (f'\033 [36mLLM Response : { response } \033 [0m ' )
135
135
return response
136
136
except Exception as e :
137
137
app_logger .info (f'{ __file__ } : { e } ' )
Original file line number Diff line number Diff line change @@ -36,7 +36,7 @@ def create_doc_gpt(
36
36
docGPT .llm = llm_model
37
37
agent_ .llm = llm_model
38
38
39
- docGPT .create_qa_chain (chain_type = 'refine' )
39
+ docGPT .create_qa_chain (chain_type = 'refine' , verbose = False )
40
40
docGPT_tool = agent_ .create_doc_chat (docGPT )
41
41
calculate_tool = agent_ .get_calculate_chain
42
42
llm_tool = agent_ .create_llm_chain ()
@@ -58,7 +58,7 @@ def create_doc_gpt(
58
58
# Use gpt4free llm model without agent
59
59
llm_model = GPT4Free (provider = g4f_provider )
60
60
docGPT .llm = llm_model
61
- docGPT .create_qa_chain (chain_type = 'refine' )
61
+ docGPT .create_qa_chain (chain_type = 'refine' , verbose = False )
62
62
return docGPT
63
63
except Exception as e :
64
64
module_logger .info (f'{ __file__ } : { e } ' )
Original file line number Diff line number Diff line change @@ -164,13 +164,14 @@ def _embeddings(self):
164
164
165
165
def create_qa_chain (
166
166
self ,
167
- chain_type : str = 'stuff' ,
167
+ chain_type : str = 'stuff' ,
168
+ verbose : bool = True
168
169
) -> BaseQaChain :
169
170
# TODO: Bug helper
170
171
self ._helper_prompt (chain_type )
171
172
chain_type_kwargs = {
172
173
'question_prompt' : self .prompt ,
173
- 'verbose' : True ,
174
+ 'verbose' : verbose ,
174
175
'refine_prompt' : self .refine_prompt
175
176
}
176
177
@@ -233,8 +234,8 @@ def _call(
233
234
run_manager : Optional [CallbackManagerForLLMRun ] = None ,
234
235
) -> str :
235
236
try :
236
- print (f'Promopt : { prompt } ' )
237
- print (f'Provider : { self .PROVIDER_MAPPING [self .provider ]} ' )
237
+ # print(f'\033[36mPromopt : {prompt}\033[0m ')
238
+ print (f'\033 [36mProvider : { self .PROVIDER_MAPPING [self .provider ]} \033 [0m ' )
238
239
return g4f .ChatCompletion .create (
239
240
model = "gpt-3.5-turbo" ,
240
241
messages = [{"role" : "user" , "content" : prompt }],
You can’t perform that action at this time.
0 commit comments