LoRA integration with llama-export-lora.exe results in blank/gibberish output #10911
Unanswered
AkiMatsushita
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm experiencing an issue when integrating a LoRA model with a base model using
llama-export-lora.exe
. After merging the LoRA into the base model's gguf file, the resulting model either produces blank output or gibberish (e.g., "---------") regardless of the prompt used.Steps to reproduce:
peft
library with the following settings:convert_lora_to_gguf.py
with the--outtype f16
option.Command:
python convert_lora_to_gguf.py --base ./base_model --outfile ./lora_model.gguf --outtype f16 ./lora_model
llama-export-lora.exe
.Command:
llama-export-lora.exe -m base_model.gguf --lora lora_model.gguf -o merged_model.gguf
Command:
./main -m merged_model.gguf -p "Test prompt"
Expected behavior:
The merged model should generate text based on the given prompt, incorporating the knowledge learned by the LoRA.
Actual behavior:
The merged model produces either blank output or gibberish (e.g., "---------"). This happens even with prompts unrelated to the training data.
Environment:
Additional context:
transformers
andpeft
.Thank you for your time and assistance.
Beta Was this translation helpful? Give feedback.
All reactions