GPT throughput (v.s. Megatron-LM) #2842
Replies: 3 comments 1 reply
-
Hi @yurishin929 Titans is our temporary model zoo because some complicate parallel strategies may require users to modify the model. We offer some mainstream models for users in Titans. |
Beta Was this translation helpful? Give feedback.
-
Hi @binmakeswell ,
Thanks in advance! |
Beta Was this translation helpful? Give feedback.
-
Hi @Agoniii Using Gemini with |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
In README, when using GPT-3 model, colossal-ai shows better performance(sec/iter and Thoughput) than Megatron-LM. and Thoughput for GPT-2, Bert too.
so my question is, for these performances, did you guys use Titans?
especially in Colossal-AI for GPT-3 Table, second one that shows 4.99 throuput.
I'm using Gemini and Megatron-LM for GPT2-medium. and Gemini has worse throughput than Megatron-LM(but better memory efficient). Does Titans show better thoughput(and lower sec/iter) than Gemini and Megatron? Thank you.
Beta Was this translation helpful? Give feedback.
All reactions