-
Notifications
You must be signed in to change notification settings - Fork 67
Closed
Labels
questionFurther information is requestedFurther information is requested
Description
Compared with llama.cpp, does tmac lose precision when running quantized models, or does it give the same results? I am running qwen1.5 4bit(https://huggingface.co/Qwen/Qwen1.5-4B-Chat-GPTQ-Int4) now, and I found that the answers given by the model are sometimes wrong, especially in English. like this:
Metadata
Metadata
Assignees
Labels
questionFurther information is requestedFurther information is requested