Convert a simple custom transformer to GGUF format #8232
Replies: 1 comment
-
This PR does it for Phi-models. Is it a good starting point? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello!
I would be very grateful if someone could tell me if what I want is possible, and if so, share some tips on how to achieve it.
Suppose I have a simple transformer model, similar to GPT, with several decoder layers. The model is structured like this:
Is there a way to convert this model to GGUF and use it for inference? I understand that for the model to be converted to GGUF and used for inference, its architecture needs to be supported by the convert-hf-to-gguf.py script. Given that the model is quite simple, maybe it is not so hard to do.
As for the tokenizer: it is a fixed mapping from words to numbers (about 240 words in total, like a simple dict)
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions