-
Notifications
You must be signed in to change notification settings - Fork 6.2k
Qwen3 not stopping generation after lora finetuning #7943
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
You can try |
I have tried the template: default for Qwen 3, and it is still the same. But if I change the model to LLama 3b base, then it's working fine. |
maybe the eos token was wrong in qwen3 base model. You can use |
遇到了一样的问题 |
Try setting the eos token in the tokenizer config to |
It worked! Thanks! |
Reminder
System Info
LLaMA-Factory.git@6a584b40928fb6d69e22c1403db226eb04358a30#egg=llamafactory
Python 3.11.12
Ubunutu 22.04
Nvidia A100 80GB
Reproduction
I'm training Qwen/Qwen3-4B-Base with language adapters. For SFT full parameter fine-tuning, the generation is coming out as expected. But when I train the same data and model with Lora, the generation is not stopping. Here is my train config
Others
No response
The text was updated successfully, but these errors were encountered: