How to convert FP16 gguf model to 4bit or 5 bit gguf model #8108
Answered
by
cshamis
RakshitAralimatti
asked this question in
Q&A
-
I am able to convert the models to fp16 gguf models but facing issue in converting them further in 4bit or 5bit can any one help me to achieve this? |
Beta Was this translation helpful? Give feedback.
Answered by
cshamis
Jul 2, 2024
Replies: 2 comments
-
the above command works for me? |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
RakshitAralimatti
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
% llama-quantize ../models/model_f16.gguf ../models/model_Q4_K_M Q4_K_M