-
Notifications
You must be signed in to change notification settings - Fork 110
fix: Input batch size overflow vulnerability #434
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
@@ -1076,7 +1076,7 @@ InferenceRequest::Normalize() | |||
|
|||
// Make sure request batch-size doesn't exceed what is supported by | |||
// the model. | |||
if ((int)batch_size_ > model_config.max_batch_size()) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
batch_size_
is uint32_t
type.
No HTTP or gRPC client test because input size is gigantic. The minimum input byte size to validate, assuming bool input with shape [-1,1], is 2^31 bytes or 2 GB. In gRPC client, the maximum byte size allowed is exactly 2^31-1 bytes.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
What does the PR do?
ValidateBytesInputs
parameters as it's not used.Checklist
<commit_type>: <Title>
Commit Type:
Check the conventional commit type
box here and add the label to the github PR.
Related PRs:
triton-inference-server/server#8165
Where should the reviewer start?
Test plan:
L0_input_validation--base
27387023
Caveats:
Background
Related Issues: (use one of the action keywords Closes / Fixes / Resolves / Relates to)