Skip to content

Qualcomm-AI-research/SHiRA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 

Repository files navigation

Sparse High Rank Adapters (NeurIPS 2024)

Sparse High Rank Adapters or SHiRA is an alternate type of adapter and has been found to have significant advantages over the low rank adapters. Specifically, SHiRA achieves better accuracy than LoRA for a variety of vision and language tasks. It also offers simpler and higher quality multi-adapter fusion by significantly reducing concept loss, a common problem faced by low rank adapters. SHiRA directly finetunes a small number of the base model’s parameters to finetune the model on any adaptation task.

The code for SHiRA is now included in the official PEFT Library. An example to run SHiRA is present at PEFT Github.

SHiRA Summary

If you find our work useful, please consider citing our work:

@inproceedings{bhardwaj2024sparse,
  title={Sparse high rank adapters},
  author={Bhardwaj, Kartikeya and Pandey, Nilesh Prasad and Priyadarshi, Sweta and Ganapathy, Viswanath and Kadambi, Shreya and Esteves, Rafael and Borse, Shubhankar and Whatmough, Paul and Garrepalli, Risheek and Van Baalen, Mart and others},
  booktitle={Proceedings of the 38th International Conference on Neural Information Processing Systems},
  pages={13685--13715},
  year={2024}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published