Skip to content

Quantizing/converting Llama 3.1 on latest master results in an unloadable model. Any thing I'm missing? #9077

Closed Answered by objectivecosta
objectivecosta asked this question in Q&A
Discussion options

You must be logged in to vote

Updating for whomever needs it: The llama-cli and other binaries in the root folder of the project are not symlinks to the recently built ones. So the llama-cli will never match with the one you just built after running the build commands.

Just build it yourself and run it from the build folder and it should work with the quantized models.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by objectivecosta
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant