API Cost Discussion #1527
Replies: 7 comments 6 replies
-
You could also try to use Cloudflare's AI Gateway here. |
Beta Was this translation helpful? Give feedback.
-
Thank-you, seems to be working very well!!
Ideally should be an environment variable but hey, just testing. Thankyou again for your great idea. |
Beta Was this translation helpful? Give feedback.
-
Hi, I re-opened this for further ideas and discussions.
|
Beta Was this translation helpful? Give feedback.
-
Update: by using the API gateway I was able to identify and hopefully address the ROOT cause. |
Beta Was this translation helpful? Give feedback.
-
@spammenotinoz Have you looked into hosted versions of Llama or Mistral? They are open source so its likely to be way cheaper than GPT. Also, Groq hosts them and they are insanely fast too. I see @mckaywrigley just added groq to the models as well! |
Beta Was this translation helpful? Give feedback.
-
The code change appears to have had a big impact on the cost\token count. |
Beta Was this translation helpful? Give feedback.
-
Anyone interested in GPT-4\GPT-4 Turbo cost reduction and better outputs when attachments are used, please see this PULL request from |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I enjoy this interface and the GPT4 Turbo model over ChatGPT, however even with low GPT-4 Turbo usage, my API costs are high ($3-6 per day), so is a lot more expensive for myself than ChatGPT. Which is great for this project as it offers easy access to cheaper models.
Curious though, even for low usage, would there be any benefit it deploying an API cache, something like "https://github.com/zilliztech/GPTCache"
Or are there any services that offer lower cost access to GPT-4 via caching, other means?
Beta Was this translation helpful? Give feedback.
All reactions