This repository contains code to train LLM with diverse PEFT techniques with custom datasets.
The dataset was sourced from kaggle.
https://www.kaggle.com/datasets/nelgiriyewithana/emotions/data
|
|
Fig 1. Emotions distribution in the dataset before and after undersampling.
Model | Accuracy | Precision | Recall | F1 | Matthews Correlation |
Training duration |
---|---|---|---|---|---|---|
ModernBERT-base | 0.94717 | 0.951624 | 0.94717 | 0.94786 | 0.93579 | 2:30:55 |
OPT-350m | 0.94708 | 0.949574 | 0.94708 | 0.94670 | 0.93545 | 2:27:34 |
RoBERTa | 0.94438 | 0.949431 | 0.94438 | 0.94505 | 0.93248 | 1:04:35 |
-
RoBERTa-LoRA vs OPT-350m-LoRA, pvalue: 0.1539
-
RoBERTa-LoRA vs ModernBERT-LoRA, pvalue: 0.8775
-
ModernBERT-LoRA vs OPT-350m-LoRA, pvalue: 0.2053
We failed to reject the
Weights for sequence classification are available on Hugging Face.