Skip to content

AISec-VoiceFingerprinting/WhisperVoiceTrace-A-Comprehensive-Analysis-of-Voice-Command-Fingerprinting

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

WhisperVoiceTrace-A-Comprehensive-Analysis-of-Voice-Command-Fingerprinting

⚠️ Experimental - PLEASE BE CAREFUL. Intended for reasearch purposes only.⚠️ This repository contains code and data of the paper WhisperVoiceTrace: A Comprehensive Analysis of Voice Command Fingerprinting

Installation

pip3 install -r requirements.txt
sudo apt install graphviz

Dataset

The WhiVo Alexa Dataset referenced in the paper can be accessed through the link provided : https://drive.google.com/drive/folders/1Bx3gxMC02GOXviDiIEdPvxxcqoCWkvb6?usp=share_link
dataset1 : data_mon_300_final.npz - monitored set consisting of 300 instances per class.
dataset2 : data_mon_1000_final.npz - monitored set consisting of almost 1000 instances per class.
dataset3 : data_unmon_final.npz - unmonitored set consisting of almost 300 instances per class.
dataset4 : data_location - voice command traces by configuring various locations, including US, UK, Germany, India, and Canada, on five Alexa devices

The WhiVo Google Dataset referenced in the paper can be accessed through the link provided : https://drive.google.com/drive/folders/1UMYmwv4INdThN4s9c1mFNvcPEyUSpTlQ?usp=sharing
The SHAME dataset referenced in the paper can be accessed through the link provided : https://drive.google.com/file/d/1K19SDZ3IdvAv_0rK6mG9d8WTpHg85gzV/view?usp=sharing
The DeepVC dataset referenced in the paper can be accessed through the link provided : https://drive.google.com/drive/folders/1l-fSX9VdZH5kF9z7gm82xgYX5ca0kRI0?usp=sharing

The information regarding commands used for category fingerprinting, location, dynamic/static, and human experiments can be found in the paper (Table 8, Table 9) and the provided link(https://docs.google.com/spreadsheets/d/15XVeFjGMaWQU9f6e-6OnUmUjQ9kS2anprYcihniMr6I/edit#gid=0). You can check the command number and voice_name from the array corresponding to 'label_name' in the npz file. Using the command number provided within the link, you can also retrieve the content of the respective command.

Usage

  1. Load Data and Convert the simple features to WhiVo features.
python3 Data_npz.py
  1. Train models through VAV1, VAV2, Size features and save the pre-trained weights.
python3 run_VAV1.py --load-VAV1 <file path of VAV1.pkl> --load-classes <file path of y.pkl>
python3 run_VAV2.py --load-VAV2 <file path of VAV2.pkl> --load-classes <file path of y.pkl>
python3 run_Size.py --load-Size <file path of Size.pkl> --load-classes <file path of y.pkl>
  1. Train and Test Ensemble model.
python3 run.py --load-VAV2 <file path of VAV2.pkl> --load-Size <file path of Size.pkl> --load-VAV1 <file path of VAV1.pkl> --load-classes <file path of y.pkl>
                               --load-weights-VAV2 <file path of pre-trained weight of VAV2>
                               --load-weights-Size <file path of pre-trained weight of Size>
                               --load-weights-VAV1 <file path of pre-trained weight of VAV1>

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •