High Token Usage and Model Limits: Best Practices & Feature Requests #451
Replies: 2 comments 1 reply
-
Hey @a-chumagin thanks so much for this thoughtful discussion. I'll have a think about some of these feature requests and get back to you. Do you have any public examples of an overuse of context finding I can check out? Best, |
Beta Was this translation helpful? Give feedback.
-
Hi @mattzcarey ,
These features would make it much easier to manage costs and stay within provider limits, especially on platforms like Azure OpenAI. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
High Token Usage and Model Limits: Best Practices & Feature Requests
I’m a user of Shippie and I’ve been running into issues with very high token usage during reviews, which has led to hitting model usage limits (especially on Azure OpenAI) and increased costs. I’d like to open a discussion on best practices for reducing token usage and controlling model limits in Shippie, and to share what I’ve already tried.
My Situation
Efforts I’ve Already Made
Custom Instructions:
I tried using
--customInstructions
to tell the agent to only use certain tools (e.g.,read_diff
,thinking
,suggest_change
,submit_summary
) and avoid others. However, this is not strictly enforced and the agent sometimes ignores it.File Exclusion:
I used the
--ignore
flag to exclude as many files and directories as possible from the review.Lowered maxSteps:
I set
"shippieMaxSteps": 20
to reduce the number of agentic steps.Minimized rules/docs:
I tried to reduce the size of project rules and documentation files that are included in the prompt. But this loses one of the advantages of Shippie—it’s cool when you can add custom rules!
Cheaper models:
I used less expensive models (like GPT-4o) for reviews, but then I faced a decrease in review quality.
Despite these efforts, I’ve noticed that:
Questions & Feature Requests
--maxPromptTokens
or similar flag to cap the prompt/context size?I believe these features would help a lot of users who are concerned about cost and provider limits, especially on platforms like Azure OpenAI/OpenAI compability platforms .
Thanks for your work on Shippie! I’d love to hear any advice, and I’m happy to help test or contribute if these features are on the roadmap.
Beta Was this translation helpful? Give feedback.
All reactions