Example Code and instructions to reproduce the Single GPU optimization demo on GPT2 shown in the repo main page #5259
Unanswered
LCChen
asked this question in
Community | Q&A
Replies: 1 comment
-
Some update : I have been able to run 1.2B model (gpt_xl) in examples/language/gpt/gemini directory using two GPUs. But I never be able to run gpt2-10b model using the same example code. Here is the configuration : ( I have minimize the batch to 1 to save the memory) root@a370836de98a:/workspace/ColossalAI/examples/language/gpt/gemini# bash run_gemini.sh
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Dear experts,
I have a two RTX3090 cards locally, and is excited about the data shown in the picture bellow (single GPU could train 18B model). I would like to try it locally myself on the same demo and help me understand the mechanism of the Colossal AI better. Can I get your helps to find the example code and instructions to run the 18B GPT2 model on one GPUs (3090) ?
Thank u a lot for the help.

Beta Was this translation helpful? Give feedback.
All reactions