Skip to content

Medical Language Model fine-tuned using pretraining, instruction tuning, and Direct Preference Optimization (DPO). Progresses from general medical knowledge to specific instruction following, with experiments in preference alignment for improved medical text generation and understanding.

Notifications You must be signed in to change notification settings

samadon1/LLM-From-Scratch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Medical LLM Project

A Language Model fine-tuned for medical applications, progressing from pretraining to instruction fine-tuning and Direct Preference Optimization (DPO).

Datasets

  1. Pretraining: Medical Text Dataset (Kaggle)
  2. Fine-tuning: PMC LLaMA Instructions (Hugging Face)

Project Stages

  1. Pretraining

    • Custom GPT model on medical texts
  2. Instruction Fine-tuning

    • Used LitGPT for LoRA fine-tuning on instruction dataset
  3. Direct Preference Optimization (DPO)

    • Generated variants using fine-tuned model
    • Created preference pairs based on Levenshtein distance

Key Features

  • Customized for medical domain
  • Progression from general language model to instruction-following
  • Experiment with preference optimization

Future Work

  • Larger medical datasets
  • Advanced DPO techniques
  • Multi-task learning in medical domain
  • Benchmark evaluation:
    • Compare against established medical NLP models
    • Evaluate on standardized medical QA datasets
    • Assess performance on clinical decision support tasks

About

Medical Language Model fine-tuned using pretraining, instruction tuning, and Direct Preference Optimization (DPO). Progresses from general medical knowledge to specific instruction following, with experiments in preference alignment for improved medical text generation and understanding.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published