Skip to content

ObsisMc/weldingLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Welding LLM

This project further pre-trained the Mistral model and fine-tuned it with instructions using LoRA, based on five welding reference textbooks, aiming to develop a large language model specialized in the welding domain.

  • model: unsloth/Mistral-Nemo-Base-2407
  • Parameter-Efficient Fine-Tuning : LoRA
  • Hardware: Colab A100, 40G

Install

For data preparation,

pip install langchain

Model training will run on Colab so you don't need to install dependencies locally.

Data Preparation

The preprocessed data is here, which contains training and evaluation data.

If you want to use the shared evaluation data, which only contains multiple-choice questions, please see eval/README.md before using it.

If you want to prepare your own training data, please see preprocess/README.md for more details. The data will be generated in data folder and the data_example folder is provided for your reference.

Finetuning

See finetune/README.md for more details.

Evaluation

This project uses a dataset with multiple-choice questions to evaluate the model as mentioned in Data Preparation section.

The initial result is here.

Future Improvement

  • Higher-Quality training and evaluation data
  • Larger LLM / Train more parameters

About

A LLM in the welding domain

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published