From eb8c6798d4c36d19dcf6095cd9c5c028e295f9d6 Mon Sep 17 00:00:00 2001 From: ArpitGaur Date: Mon, 13 May 2024 09:58:09 +0100 Subject: [PATCH 1/3] starting a branch for this pairing task --- infra/main.bicep | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/infra/main.bicep b/infra/main.bicep index 64d23b87d..ff2d307b9 100644 --- a/infra/main.bicep +++ b/infra/main.bicep @@ -119,7 +119,7 @@ param azureOpenAIVisionModelVersion string = 'vision-preview' @description('Azure OpenAI Vision Model Capacity - See here for more info https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/quota') param azureOpenAIVisionModelCapacity int = 10 -@description('Orchestration strategy: openai_function or langchain str. If you use a old version of turbo (0301), plese select langchain') +@description('Orchestration strategy: openai_function or langchain str. If you use a old version of turbo (0301), please select langchain') @allowed([ 'openai_function' 'langchain' From 38f6612378a2641a19ff14d584ea4dc7d678a49d Mon Sep 17 00:00:00 2001 From: ArpitGaur Date: Mon, 13 May 2024 10:45:32 +0100 Subject: [PATCH 2/3] removing TODOs --- code/backend/batch/utilities/plugins/ChatPlugin.py | 2 -- code/backend/batch/utilities/plugins/PostAnsweringPlugin.py | 1 - 2 files changed, 3 deletions(-) diff --git a/code/backend/batch/utilities/plugins/ChatPlugin.py b/code/backend/batch/utilities/plugins/ChatPlugin.py index eb3e79454..2d672537a 100644 --- a/code/backend/batch/utilities/plugins/ChatPlugin.py +++ b/code/backend/batch/utilities/plugins/ChatPlugin.py @@ -21,7 +21,6 @@ def search_documents( str, "A standalone question, converted from the chat history" ], ) -> Answer: - # TODO: Use Semantic Kernel to call LLM return QuestionAnswerTool().answer_question( question=question, chat_history=self.chat_history ) @@ -37,7 +36,6 @@ def text_processing( "The operation to be performed on the text. Like Translate to Italian, Summarize, Paraphrase, etc. If a language is specified, return that as part of the operation. Preserve the operation name in the user language.", ], ) -> Answer: - # TODO: Use Semantic Kernel to call LLM return TextProcessingTool().answer_question( question=self.question, chat_history=self.chat_history, diff --git a/code/backend/batch/utilities/plugins/PostAnsweringPlugin.py b/code/backend/batch/utilities/plugins/PostAnsweringPlugin.py index b6aa6128a..1c5d87378 100644 --- a/code/backend/batch/utilities/plugins/PostAnsweringPlugin.py +++ b/code/backend/batch/utilities/plugins/PostAnsweringPlugin.py @@ -8,5 +8,4 @@ class PostAnsweringPlugin: @kernel_function(description="Run post answering prompt to validate the answer.") def validate_answer(self, arguments: KernelArguments) -> Answer: - # TODO: Use Semantic Kernel to call LLM return PostPromptTool().validate_answer(arguments["answer"]) From 4f16864fbdd355b6cc2f1a083bb33221833aadf6 Mon Sep 17 00:00:00 2001 From: ArpitGaur Date: Mon, 13 May 2024 12:06:03 +0100 Subject: [PATCH 3/3] update semantic_kernel in the allowed list. --- infra/main.bicep | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/infra/main.bicep b/infra/main.bicep index ff2d307b9..f19c80f28 100644 --- a/infra/main.bicep +++ b/infra/main.bicep @@ -119,9 +119,10 @@ param azureOpenAIVisionModelVersion string = 'vision-preview' @description('Azure OpenAI Vision Model Capacity - See here for more info https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/quota') param azureOpenAIVisionModelCapacity int = 10 -@description('Orchestration strategy: openai_function or langchain str. If you use a old version of turbo (0301), please select langchain') +@description('Orchestration strategy: openai_function or semantic_kernel or langchain str. If you use a old version of turbo (0301), please select langchain') @allowed([ 'openai_function' + 'semantic_kernel' 'langchain' ]) param orchestrationStrategy string = 'openai_function'