🔧 The code is currently being organized and will be released soon. Thanks for your attention!
Traffic sign cross-regional recognition and results. (a) introduces the main content of this paper, fine-tuning TSCLIP on specific traffic sign datasets, and then performing recognition on other worldwide regions. (b) shows our TSCLIP model is far superior to the classic model and exceeds the mainstream CLIP fine-tuning scheme.
Four representative traffic signs (No Overtaking, No Parking, No Pedestrians, and Stop).
Robust fine-tuning framework for TSCLIP model. (a) shows the contrastive language-image training process of TSCLIP with traffic sign prompts (b) shows the Wise-FT scheme for weight ensembling of the CLIP model (c) shows our proposed Adaptive Dynamic Weight Ensembling (ADWE) scheme.
We constructed the CRTS dataset with over 190,000 images, integrating traffic sign from 10 regions. The CRTS standardizes 46 commonly used categories with unified labels for consistent cross-regional recognition.
The download link for the CRTS dataset is HERE.
If you find TSCLIP useful in your research or applications, please consider giving us a star 🌟 and citing it.
@misc{zhao2024tscliprobustclipfinetuning,
title={TSCLIP: Robust CLIP Fine-Tuning for Worldwide Cross-Regional Traffic Sign Recognition},
author={Guoyang Zhao and Fulong Ma and Weiqing Qi and Chenguang Zhang and Yuxuan Liu and Ming Liu and Jun Ma},
year={2024},
eprint={2409.15077},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2409.15077},
}