Skip to content

Repository containing implementation for co-training large (e.g., T0) and smaller (e.g., BERT) language models to enhance few-shot performance

License

Notifications You must be signed in to change notification settings

lopeve/colearnt-prompting

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Enhanced Co-training for Large Language Models

This repository contains the implementation for our ICML 2022 paper Co-training Improves Prompt-based Learning for Large Language Models and subsequent advancements, including tuning methodologies based on T-Few.

The code is instrumental for:

  • Enhancing the zero-shot and few-shot performance of large language models
  • Distilling large models like GPT-3 and T0 into compact task-specific models.

We sucesfully built many parts of this repository on top of the outstanding T-Few repository.

If you find this code useful, please consider citing our paper:

@inproceedings{lang2022co,
  title={Co-training improves prompt-based learning for large language models},
  author={Lang, Hunter and Agrawal, Monica N and Kim, Yoon and Sontag, David},
  booktitle={International Conference on Machine Learning},
  pages={11985--12003},
  year={2022},
  organization={PMLR}
}

Setup, usage, model training, result reproduction, and method of application to your data set are outlined in detail in the original README content.

Note:

This is the new home for the project, under the new owner lopeve, for any references or author contacts please see the original paper and repository maintained by clinicalml.

About

Repository containing implementation for co-training large (e.g., T0) and smaller (e.g., BERT) language models to enhance few-shot performance

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages