Skip to content

Commit adc7b63

Browse files
feat: Update Solution to Leverage Your Own Data
2 parents 664faf2 + f7e3fd5 commit adc7b63

File tree

7 files changed

+830
-2
lines changed

7 files changed

+830
-2
lines changed

documents/CustomizeData.md

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,23 @@
1+
## Customize the solution with your own data
2+
3+
If you would like to update the solution to leverage your own data please follow the steps below.
4+
> Note: you will need to complete the deployment steps [here](./DeploymentGuide.md) before proceeding.
5+
6+
## Prerequisites:
7+
1. Your data will need to be in JSON or wav format with the file name formated prefixed with "convo" then a GUID followed by a timestamp. For more examples of the data format, please review the sample transcripts and audio data included [here](/infra/data/)
8+
* Example: convo_32e38683-bbf7-407e-a541-09b37b77921d_2024-12-07 04%3A00%3A00
9+
10+
11+
1. Navigate to the storage account in the resource group you are using for this solution.
12+
2. Open the `data` container
13+
3. If you have audio files, upload them to `custom_audiodata` folder. If you have call transcript files, upload them to `custom_transcripts` folder.
14+
4. Navigate to the terminal and run the `run_process_data_script.sh` to process the new data into the solution with the following commands.
15+
```shell
16+
cd infra/scripts
17+
18+
az login
19+
20+
bash run_process_data_script.sh resourcegroupname_param
21+
```
22+
a. resourcegroupname_param - the name of the resource group.
23+

documents/DeploymentGuide.md

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -210,4 +210,8 @@ To help you get started, here are some **Sample Questions** you can ask in the a
210210
- Give a summary of billing issues
211211
- When customers call in about unexpected charges, what types of charges are they seeing?
212212
213-
These questions serve as a great starting point to explore insights from the data.
213+
These questions serve as a great starting point to explore insights from the data.
214+
215+
## Next Steps:
216+
Now that you've completed your deployment, you can start using the solution. Try out these things to start getting familiar with the capabilities:
217+
* [Customize the solution](./CustomizeData.md) with your own data

infra/process_data_scripts.bicep

Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
param solutionLocation string
2+
param keyVaultName string
3+
param identity string
4+
5+
var baseUrl = 'https://raw.githubusercontent.com/microsoft/Conversation-Knowledge-Mining-Solution-Accelerator/main/'
6+
7+
resource process_data_scripts 'Microsoft.Resources/deploymentScripts@2020-10-01' = {
8+
kind:'AzureCLI'
9+
name: 'process_data_scripts'
10+
location: solutionLocation // Replace with your desired location
11+
identity: {
12+
type: 'UserAssigned'
13+
userAssignedIdentities: {
14+
'${identity}' : {}
15+
}
16+
}
17+
properties: {
18+
azCliVersion: '2.52.0'
19+
primaryScriptUri: '${baseUrl}infra/scripts/process_data_scripts.sh'
20+
arguments: '${baseUrl} ${keyVaultName}' // Specify any arguments for the script
21+
timeout: 'PT1H' // Specify the desired timeout duration
22+
retentionInterval: 'PT1H' // Specify the desired retention interval
23+
cleanupPreference:'OnSuccess'
24+
}
25+
}

infra/scripts/copy_kb_files.sh

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,4 +32,6 @@ echo "Script Started"
3232
az login --identity --client-id ${managedIdentityClientId}
3333
# Using az storage blob upload-batch to upload files with managed identity authentication, as the az storage fs directory upload command is not working with managed identity authentication.
3434
az storage blob upload-batch --account-name "$storageAccount" --destination data/"$extractedFolder1" --source /mnt/azscripts/azscriptinput/"$extractedFolder1" --auth-mode login --pattern '*' --overwrite
35-
az storage blob upload-batch --account-name "$storageAccount" --destination data/"$extractedFolder2" --source /mnt/azscripts/azscriptinput/"$extractedFolder2" --auth-mode login --pattern '*' --overwrite
35+
az storage blob upload-batch --account-name "$storageAccount" --destination data/"$extractedFolder2" --source /mnt/azscripts/azscriptinput/"$extractedFolder2" --auth-mode login --pattern '*' --overwrite
36+
az storage fs directory create --account-name "$storageAccount" --file-system data --name custom_audiodata --auth-mode login
37+
az storage fs directory create --account-name "$storageAccount" --file-system data --name custom_transcripts --auth-mode login

0 commit comments

Comments
 (0)