Skip to content

Where can I download the LLaMa model weights? #4576

Answered by supportend
jzry asked this question in Q&A
Discussion options

You must be logged in to vote

I cloned the llama.cpp source with git, build it with make and downloaded GGUF-Files of the models. When i use the exact prompt syntax, the prompt was trained with, it worked.

Good source for GGUF-files: https://huggingface.co/TheBloke

Sure, when you use a graphic card, perhaps you have to enable something, to make it work.

65B 30B 13B 7B vocab.json is not a command, you have to execute. I have my models in two folders and use them this way (CPU only):

./main -t 6 -m ~/Downloads/models/mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf -c 8192 --temp 0.7 --repeat_penalty 1.1 --log-disable -n -1 -p "<s>[INST] Write a short text about UPX. [/INST]"

Take care of useable RAM and RAM consumption and adjus…

Replies: 3 comments 2 replies

Comment options

You must be logged in to vote
0 replies
Answer selected by jzry
Comment options

You must be logged in to vote
1 reply
@jzry
Comment options

Comment options

You must be logged in to vote
1 reply
@phymbert
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
5 participants