Replies: 4 comments 3 replies
-
Did you clone llama.cpp? Without |
Beta Was this translation helpful? Give feedback.
-
Looks like the |
Beta Was this translation helpful? Give feedback.
-
It's 2025 now is there anyone found a consistent solution for converting the |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
Hope I can get some pointers to get me in the right direction.
Background:
I understand that meta's llama3 model just gone GA not long ago. Have been using llama.cpp (it works great, thank you!) on gguf format of llama2 model. Wanted to try out the llama3, so conversion is required to obtain the gguf format of the model.
Progress:
Have used (convert_llama_weights_to_hf.py) to convert from Meta format (.pth) to hf (.safetensors, why is it not pytorch bin format??) format successfully. I understand from README that convert.py does not work for llama3 model at the moment, have to use 'convert-hf-to-gguf.py' instead.
When I run 'convert-hf-to-gguf.py' something like following, it throws an error about MODEL_ARCH. What did I miss?
Don't think I am using ORION.
--------- console output --------------->
[userid]$ python convert-hf-to-gguf.py hf-model-path/ --outfile output-file-path --outtype bf16
Traceback (most recent call last):
File "/home/myhome/convert-hf-to-gguf.py", line 787, in
class OrionModel(Model):
File "/home/myhome/convert-hf-to-gguf.py", line 788, in OrionModel
model_arch = gguf.MODEL_ARCH.ORION
^^^^^^^^^^^^^^^^^^^^^
AttributeError: type object 'MODEL_ARCH' has no attribute 'ORION'
I have read "Tutorial: How to convert HuggingFace model to GGUF format #2948", couldn't find anything that works.
Beta Was this translation helpful? Give feedback.
All reactions