Replies: 1 comment 2 replies
-
As far as I understood: Gatt is a special algorithm that was used when LLaMA 2 Chat was fine tuned. The GAtt is not something that requires a change in how inference is done. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Chapter 3.3 of "Llama 2: Open Foundation and Fine-Tuned Chat Models" talks about using GAtt to avoid having to repeat instructions in multi-turn dialogue. Is this something that would need to be implemented in llama.cpp, or does it already exist?
Beta Was this translation helpful? Give feedback.
All reactions