Description
I have noticed a very weird change when I wanted to make use of streaming. Before I was not, and basically all conversation models tended to start their message with an emoji. The reasons why the models are so fixated on starting the message that way is unclear to me, but the emojis clearly made sense and represented fitting emotions and such.
When I now tried to integrate streaming, I noticed the first few output chunks are empty, and now I don't see the model use emojis at all anymore. Certainly not at the start, only superfluous spaces. And less spaces than there were superfluous generation chunks before the text.
This leads me to the suspicion, that the chunking for streaming is breaking up unicode characters that generate from multiple tokens and can not be converted from byte buffer to string individually. And that causes a broken result, since chaining the outputs does not give you back the correct unicode symbol.
Might there be something to this? Or am I doing something wrong?
Edit: I am not using "echo", I imagine that would somewhat fix it. Probably still shouldn't have this effect one way or the other.