Skip to content

Wb-az/peft_lora_opt_llm_emotion_analysis

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 

Repository files navigation

LLM's PEFT

This repository contains code to train LLM with diverse PEFT techniques with custom datasets.

Task 1 - Emotion Classification

1.1 Dataset

The dataset was sourced from kaggle.

https://www.kaggle.com/datasets/nelgiriyewithana/emotions/data

Fig 1. Emotions distribution in the dataset before and after undersampling.

1.2 Results

Model Accuracy Precision Recall F1 Matthews
Correlation
Training duration
ModernBERT-base 0.94717 0.951624 0.94717 0.94786 0.93579 2:30:55
OPT-350m 0.94708 0.949574 0.94708 0.94670 0.93545 2:27:34
RoBERTa 0.94438 0.949431 0.94438 0.94505 0.93248 1:04:35

1.3 Kruskal-Wallis test

  • RoBERTa-LoRA vs OPT-350m-LoRA, pvalue: 0.1539

  • RoBERTa-LoRA vs ModernBERT-LoRA, pvalue: 0.8775

  • ModernBERT-LoRA vs OPT-350m-LoRA, pvalue: 0.2053

We failed to reject the $H_0$, indicating the models' performances have the same central tendency.

1.4 Trained adapters

Weights for sequence classification are available on Hugging Face.