Python 3.7
Pytorch 1.5.0
cuda 10.2
Clone this repository using:
git clone https://github.com/KK-xi/My_VMST.git
Create a conda environment using the environment.yml file:
conda env create -f environment.yml
To generate the voxels, we refer to the code of VoxelNet and TimoStoff.
Take N-Caltech101 as an example:
Training voxels are saved in './data/N-Caltech101/train' folder.
Testing voxels are saved in './data/N-Caltech101/test' folder.
Each sample should contains feature and coords of voxels and label.
Take N-Caltech101 as an example:
python main.py --train_dataset ./data/N-Caltech101/train/ --test_dataset ./data/N-Caltech101/test/ --arch_name VMST-Net_N-Cal --num_classes 101 --voxel_num 1024
@article{liu2023voxel,
title={Voxel-based multi-scale transformer network for event stream processing},
author={Liu, Daikun and Wang, Teng and Sun, Changyin},
journal={IEEE Transactions on Circuits and Systems for Video Technology},
volume={34},
number={4},
pages={2112--2124},
year={2023},
publisher={IEEE}
}