potential context processing levenstien culling optimization #3956
Replies: 2 comments 10 replies
-
I'm finding it hard to follow the description. Would be better if you provide a specific example. |
Beta Was this translation helpful? Give feedback.
-
Aren't they all deterministic?
If you know whether the next token beats the difference, then don't you already know the next token? Unfortunately, the description is pretty hard to follow. Are you trying to say something like given certain tokens it's basically guaranteed that other ones will follow. For example, like if you have |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I have been messing with a personal deterministic model architecture and came up with a potential context optimization though I don't know the availability of the required info in traditional models or if this is already done. This is based on in my head theory crafting I haven't tried implementing it yet.
Requirements:
Knowledge of the tokens following the previous token in the dataset
the levenstien distance for the entire context for the top scoring(lowest to highest) tokens
variable length vocabulary
Process:
find the difference in the levenstien distance between the top scoring tokens
if the next token doesn't beat this difference it means that the next token must be one of the tokens following the previous token in the dataset.
with this knowledge we can reduce the number of tokens to process to those that are long enough to produce a levenstien distance score of at least that difference.
you could further reduce the number to process by trimming the beginning and end of each token where it matches the context since those characters would not influence the levenstien distance.
(the rest is much more theory since I can't fully think through it)
if you do find tokens with a bigger levenstien distance after calculation keep track of the largest one.
my guess would be that you would be looking at the tokens that follow the token with the smallest levenstien distance that is bigger in the dataset.
Beta Was this translation helpful? Give feedback.
All reactions