Skip to content

FHR-L/VSLA-CLIP

Repository files navigation

Cross-Platform Video Person ReID: A New Benchmark Dataset and Adaptation Approach(PDF)

Installation

conda create -n vslaclip python=3.8
conda activate vslaclip
conda install pytorch==1.8.0 torchvision==0.9.0 torchaudio==0.8.0 cudatoolkit=10.2 -c pytorch
pip install yacs
pip install timm
pip install scikit-image
pip install tqdm
pip install ftfy
pip install regex

Training

For example, if you want to run for the ls-vid, you need to modify the config file to

DATASETS:
   NAMES: ('lsvid')
   ROOT_DIR: ('your_dataset_dir')
OUTPUT_DIR: 'your_output_dir'

Then, if you want to use weight of VIFI-CLIP to initialize model, you need to down the weight form link and modify config file as:

MODEL:
  VIFI_WEIGHT : 'your_dataset_dir/vifi_weight.pth'
  USE_VIFI_WEIGHT : True

If you want to run FT-CLIP (fine tune image encoder):

CUDA_VISIBLE_DEVICES=0 python train_fine_tune.py --config_file configs/ft/vit_ft.yml

if you want to run VSLA-CLIP:

CUDA_VISIBLE_DEVICES=0 python train_reidadapter.py --config_file configs/adapter/vit_adapter.yml

Evaluation

For example, if you want to test VSLA-CLIP for LS-VID

CUDA_VISIBLE_DEVICES=0 python test.py --config_file 'your_config_file' TEST.WEIGHT 'your_trained_checkpoints_path/ViT-B-16_120.pth'

Weights

Dataset LS-VID MARS iLIDS G2A
VSLA-CLIP‡ model model model model

Citation

@inproceedings{zhang2024cross,
  title={Cross-platform video person reid: A new benchmark dataset and adaptation approach},
  author={Zhang, Shizhou and Luo, Wenlong and Cheng, De and Yang, Qingchun and Ran, Lingyan and Xing, Yinghui and Zhang, Yanning},
  booktitle={European Conference on Computer Vision},
  pages={270--287},
  year={2024},
  organization={Springer}
}

Acknowledgement

Codebase from CLIP-ReID, TransReID, CLIP, and CoOp.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages