Replies: 1 comment
-
so... is it possible in such way serve model? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have followed the steps on Unsloth official notebook Alpaca + Llama-3 8b full example and finetuned a llama 3 8B model and I wanted to serve it using vllm? However it does not seem to work.
This is the command I used for serving the local model, with "/content/merged_llama3" being the directory that contains all model files:
which returns an error:
I dont think I ever need to provide quantization method as it should be written in the config file, it should be a mistake reading all those files. In addition, I did save the model and pushed it to the hub using the given codes in the Unsloth notebook?
my model files:
what went wrong?
Beta Was this translation helpful? Give feedback.
All reactions