Skip to content

Error mapping tensor using convert_hf_to_gguf.py #8521

Answered by compilade
vcc715 asked this question in Q&A
Discussion options

You must be logged in to vote

Thanks. In this case, it seems like model.embed_tokens.weight is pre-quantized, since that tensor is in U32, and it's accompanied with scales and biases.

The convert script does not yet support that, unfortunately. The model.embed_tokens.weight tensor would need to be dequantized first by applying model.embed_tokens.scales and model.embed_tokens.biases to it (not sure how exactly).

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@vcc715
Comment options

@compilade
Comment options

Answer selected by vcc715
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants