Help with model output inconsistencies? #4058
-
Can I get some help with consistency in my model outputs, not in the words, but in the structure, I feel like it works really well 50% of the time. The strange thing is I have the same prompt using Ollama with the same model and it seems to perform far better with 99% of runs working well. Command Im running:
This is the contents of the prompt file:
Sometimes I get good responses like this:
But other times I'll just get something like this:
or this:
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
The first trailing newline in a prompt is stripped off. So if you want an actual newline in the prompt, try adding two. In other words, if your prompt is:
and you want that newline, then you want the file to look like:
Also, I'm not really familiar with Ollama but you're not using the correct LLaMA2-chat instruction format. Maybe Ollama is fixing that for you somehow behind the scenes. If I remember correctly, it's supposed be like:
You can sometimes get away with violating the prompt format the model was trained on but sometimes it can make a pretty big difference to the quality of the response. |
Beta Was this translation helpful? Give feedback.
The first trailing newline in a prompt is stripped off. So if you want an actual newline in the prompt, try adding two. In other words, if your prompt is:
and you want that newline, then you want the file to look like:
Also, I'm not really familiar with Ollama but you're not using the correct LLaMA2-chat instruction format. Maybe Ollama is fixing that for you somehow behind the scenes. If I remember correctly, it's supposed be like:
You can sometimes get away with violating the prompt format the model was trained on but sometimes it can make a pretty big…