-
In here you are suggesting batch 64 for 80 GB GPU memory. Which model type (small, medium, large, xlarge) does this batch size applicable? |
Beta Was this translation helpful? Give feedback.
Answered by
titu1994
May 25, 2023
Replies: 1 comment 2 replies
-
This is for the Large model (the default in the config) with 120 M params |
Beta Was this translation helpful? Give feedback.
2 replies
Answer selected by
mehadi92
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
This is for the Large model (the default in the config) with 120 M params