Endless Responses and Imaginary Conversations in llama.cpp #4187
-
When running llama.cpp with the provided command in the terminal, the models' responses extend beyond the expected answers, creating imaginary conversations. Instead of succinctly answering questions, the output includes system information and unrelated details. This behavior persists across various models. The issue interferes with the intended use of the models. Execute the following command in the terminal:
And so on with other models I have tried, how can I fix it. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 3 replies
-
You're passing the literal escape codes in, not newlines. So a I also feel like you probably didn't paste the exact command you used to generate that output which can make it kind of help. (It's possible I'm wrong and there's something magical in |
Beta Was this translation helpful? Give feedback.
-
this site helped me with that : https://www.promptingguide.ai/ |
Beta Was this translation helpful? Give feedback.
Like this:
./main -ngl 60 -m ./models/openhermes-2.5-mistral-7b-16k.Q4_K_M.gguf --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<|im_start|>system \n Your name is Nevil. <|im_end \n <|im_start|>user \n What is your name?<|im_end|> \n <|im_start|>assistant" -r"<|im_end|>"