Replies: 1 comment 2 replies
-
Hi @Saigut, it is required to launch Triton with the script |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
The
./scripts/launch_triton_server.py
in tensorrtllm_backend repo has a parameter--world_size
, buttritonserver
don't.So when I try to launch a "tp_size=4 pp_size=1" tensorrtllm model using
tritonserver
, it will report error like mpiSize != tp_size*pp_size, due to mpiNUm is always 1.Is there any method for
tritonserver
to specifyworld_size
for tensorrtllm model?I think the most proper way maybe to set
world_size
in model repo of tensorrtllm model, like inconfig.pbtxt
, but I can't find such a option.Beta Was this translation helpful? Give feedback.
All reactions