Skip to content

This repository contains the implementation of a Reinforcement Learning with Human Feedback (RLHF) system using custom datasets. The project utilizes the trlX library for training a preference model that integrates human feedback directly into the optimization of language models.

Notifications You must be signed in to change notification settings

SJ9VRF/Reinforcement-Learning-for-Human-Feedback-RLHF

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Reinforcement Learning for Human Feedback (RLHF)

This repository contains the implementation of a Reinforcement Learning with Human Feedback (RLHF) system using custom datasets. The project utilizes the trlX library for training a preference model that integrates human feedback directly into the optimization of language models.

Screenshot_2024-08-17_at_12 25 37_AM-removebg-preview

Installation

To set up your environment to run this project, follow these steps:

  1. Clone the repository:

    git clone https://github.com/your_username/RLHF_Project.git
    cd RLHF_Project
  2. Install the required dependencies:

    pip install -r requirements.txt

Usage

To run the RLHF training process, execute the main.py script:

python main.py

About

This repository contains the implementation of a Reinforcement Learning with Human Feedback (RLHF) system using custom datasets. The project utilizes the trlX library for training a preference model that integrates human feedback directly into the optimization of language models.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages