I know this would require a new model structure, or maybe not, but hear me out. #2612
Replies: 4 comments 1 reply
-
I'll add: |
Beta Was this translation helpful? Give feedback.
-
This is a good idea, like you many others have thought of it however majority of us are just waiting for second coming of Falcon. :) |
Beta Was this translation helpful? Give feedback.
-
I think the main problem would be with sequential nature of LLM, they are currently unable to replace tokens in the middle of text, without having the entire "rough" text in the context and rewriting it each time, adding back to the context, and so on. In order to "improve" a text, gradually, the "previous" text has to be in the context, which to my understanding, uses more context then the current approach. Edit: And you can do that already, writing an agent to split your prompt into prompt for reasoning, conclusion, and final answer, passing to the output only the final answer. It's more or less how langchain works. |
Beta Was this translation helpful? Give feedback.
-
Not if the full text is not used in the context, and not if the full text
is produced as part of a final pass, via a Grammar generator function.
…On Tue, Aug 22, 2023 at 7:45 AM staviq ***@***.***> wrote:
I think the main problem would be with sequential nature of LLM, they are
currently unable to replace tokens in the middle of text, without having
the entire "rough" text in the context and rewriting it each time, adding
back to the context, and so on.
In order to "improve" a text, gradually, the "previous" text has to be in
the context, which to my understanding, uses more context then the current
approach.
—
Reply to this email directly, view it on GitHub
<#2612 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABJ5LB7PSE3YYAHMR7Z5EI3XWTAWZANCNFSM6AAAAAA3P2R4YU>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Instead of relying on attention, what if the base structure of the response were built first, which could be done quickly. Then add the filler words after the structure is complete. the 2 tasks could be done very quickly, and the base structure could extend attention to much longer, more sane responses.
for instance:
Prompt: "How would someone bake a cake from scratch?"
Thought process:
"Cake. Bake == Make. From Scratch. [INFER]"
Answering:
to make a cake from scratch. Cup Flour. 1 Egg. Cup Water. butter. cream. Mix. Oven. 350f. 30 minutes. toothpick comes clean poke center.
Before filling in the easier to quantify filler words that make the sentence easier to read for humans:
To make a cake from scratch:
Take 1 cup of flour, and mix in 1 egg, 1 cup of water, some butter (Add to taste), some cream (also to taste), and mix in a bowl until all lumps are gone. Pour into a cake pan, and gently place in a preheated oven at 350F (176.6C) for 30 minutes, or until a toothpick comes out clean when poked in the center.
Beta Was this translation helpful? Give feedback.
All reactions