use llama.cpp quantization with QAT finetune #11170
Unanswered
anyinlover
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm finetuning a 1B model with QAT using torchtune library, but I wonder is it match to llama.cpp quantization algorithm, does anyone have some experience?
Beta Was this translation helpful? Give feedback.
All reactions