-
Notifications
You must be signed in to change notification settings - Fork 12.4k
Closed as not planned
Description
I downloaded a model alpaca-30b-lora-int4
from https://huggingface.co/elinas/alpaca-30b-lora-int4/tree/main
The model is a .safetensors
in GPTQ format I think
I need to convert it to GGML .bin
so I used the script provided in llama.cpp
with the command python convert-gptq-to-ggml.py models/30B/alpaca-30b-4bit.safetensors models/30B//tokenizer.model models/30B/alpaca-30b-4bit.bin
But I get the following error
Traceback (most recent call last):
File "/big/meyer/expe/llama.cpp/convert-gptq-to-ggml.py", line 21, in <module>
model = torch.load(fname_model, map_location="cpu")
File "/big/meyer/expe/llama.cpp/.venv/lib/python3.10/site-packages/torch/serialization.py", line 815, in load
return _legacy_load(opened_file, map_location, pickle_module, **pickle_load_args)
File "/big/meyer/expe/llama.cpp/.venv/lib/python3.10/site-packages/torch/serialization.py", line 1035, in _legacy_load
raise RuntimeError("Invalid magic number; corrupt file?")
RuntimeError: Invalid magic number; corrupt file?
How to use .safetensors
models with llama.cpp
?
ericek111, Lissanro, James-E-A, rtrad89 and 4ndr3aR
Metadata
Metadata
Assignees
Labels
No labels