Skip to content

How to run LLAMA 2 70B model using llama.cpp: not working on new build #3015

Answered by KerfuffleV2
JRZS asked this question in Q&A
Discussion options

You must be logged in to vote

The model was converted to the new format gguf, but since that change, everything has broken.

The issue is the conversion, not trying to run in.

You need to specify --gqa 8 --eps 1e-5 for the GGML to GGUF conversion script. (The --gqa one is what's causing your error but using the wrong eps value will affect the quality of your output.)

Replies: 2 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@JRZS
Comment options

Answer selected by JRZS
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants