Replies: 1 comment
-
@mseddon I'm currently working on #85, which will include better support for infinite test generation and automatic context swapping. I aim to make the API as high-level as possible with good defaults, so while you can still configure more advanced parameters, using this library won't require too much technical understanding of the underlying code. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
In llama.cpp, the
main
command supports saving the prompt via the--keep
flag, which is handy for long running chat sessions. Is there a way to simulate this with the current API? After a few exchanges, it appears the context fills up and the whole system goes bananas.What I really need is access to
llama_kv_cache_seq_shift
andllama_kv_cache_seq_rm
Beta Was this translation helpful? Give feedback.
All reactions