Writing Prediction with GPT-2 model #27
Replies: 4 comments 4 replies
-
In fact, I've thought about this before, which is why I added the OpenAI Playground integration. I could imagine this with a new chatGPT model like GPT-3.5, since currently everyone has access to it via the API key. But if you do this via OpenAI's API key, you have to keep recharging, which in turn costs something. What I would like even better would be Code-LLama, but I don't know what the API key is. It might be convenient to run Code-LLama locally on the device via the application, but that would cost a lot of performance, which I don't think is great. However, one option would be to use a free API key. In my opinion, perhaps the best option would be if one could somehow integrate Microsoft Copilot or GitHub Copilot, which has always provided suggestions in VS Code. And now you can also answer that Github Copilot costs something. As a rule, there are monthly costs. But that wouldn't be the case if you verified yourself as a student. In addition, my software already has a very extensive integration with Github, which is why I consider Github Copilot almost the best. |
Beta Was this translation helpful? Give feedback.
-
I think we can try to use the Llama or GPT-2.0 fine tuned version model. It is not as good as 3.5 and is only capable of finishing and continuing the sentence(instead of Q&A), but it can be used without internet(run locally). Although it requires some bit of CPU power but the lightweight model I guess will be fine for some CPU like intel i7 10th Gen, 16Gb or so RAM. Because I was in a deep learning class in school so we made some sentence continuers like these before(not super hard just that I don't know how to add them onto the user interface). |
Beta Was this translation helpful? Give feedback.
-
Falcon7BC.py.zip Of course this is quite large as it is fine tuned to be capable like early builds of GPT3.0; another method is just changing the model.load to 'gpt2' and from transformers import gpt2, which will create a new blank model that is usable but better with finetuning. It is around 200Mb when first created, and you can allow the model to train on the content the user writes in the text-editor once they are saved(by training in another thread) and using time.sleep in the thread to decrease CPU usage power and train more slowly for older computers. |
Beta Was this translation helpful? Give feedback.
-
(and the output in terminal looks like this without the l1 l2 stuff because I altered the base code) |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Is it possible for the program to get the text that the user is on and then also show a grey(hint) word in front of the sentence so that the user can try to finish their sentence by using the AI predictions? We can use the GPT-2 model or some of the pretrained llama or alpaca models for the prediction.
(To be honest, I am only more familiar with pygame and tkinter because we use them constantly in school).
Beta Was this translation helpful? Give feedback.
All reactions