Replies: 2 comments
-
you need a GGUF v2 file format to run with the latest llama.cpp. There have been a LOT of LLM model updates in the past 7 months... perhaps there is a better and more recent model in GGUFv2? TheBloke has many already converted and quantized: https://huggingface.co/TheBloke?search_models=GGUF |
Beta Was this translation helpful? Give feedback.
-
Although they're both named The first repo is the original full unquantized model in PyTorch format. If you have the time/disk space you can download that and use the |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I have found 2 different repos for this, and i only see bin files from 7 months ago.
1- https://huggingface.co/chavinlo/gpt4-x-alpaca/tree/main
2- https://huggingface.co/anon8231489123/gpt4-x-alpaca-13b-native-4bit-128g/tree/main/gpt4-x-alpaca-13b-ggml-q4_1-from-gptq-4bit-128g
They only provide a bin file, and I can't get llama.cpp to open it so I can't interact with this model.
Any help is appreciated.
Beta Was this translation helpful? Give feedback.
All reactions