Efficient Streaming Language Models with Attention Sinks [ #3443
logikstate
started this conversation in
Ideas
Replies: 1 comment 3 replies
-
Hmm, either you misunderstood it or I did. https://github.com/mit-han-lab/streaming-llm#faq says that it doesn't actually increase context. As I understand it, this is basically #3377 - a more graceful way to do |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
New paper with example code claims huge context with minimal changes
https://github.com/mit-han-lab/streaming-llm
Beta Was this translation helpful? Give feedback.
All reactions