Convert HF format model to gguf format and the model turns very big #8322
Unanswered
SeanZhang7
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I converted the merged HF model(Codellama-7B) into gguf format with examples/convert_legacy_llama.py and it went from 14 gb to 25 gb. Is this reasonable?
And when I try to use convert_hf_to_gguf.py, I got the error message: ValueError: Duplicated key name 'tokenizer.chat_template'
Beta Was this translation helpful? Give feedback.
All reactions