Handle 413 prompt too large response from APIs like vertex gracefully #2277
ianlintner
started this conversation in
Feature Requests
Replies: 1 comment
-
+1 It's not just vertex AI (Google's models have a 1M token window, one of the largest). It's worse on Anthropic's models that have a 200K token window - trying to read one nontrivial npm lock file will blow that away every time. I have specifically changed my prompts for one project to exclude lockfiles for this reason. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
As a user if roo detects a prompt too long response rather than getting stuck in a retry loop could handle it automatically without user input.
e.g. From google vertex ai
Beta Was this translation helpful? Give feedback.
All reactions