We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 99c8489 commit c1feccdCopy full SHA for c1feccd
vllm_ascend/worker/worker_v1.py
@@ -247,9 +247,10 @@ def pin_lora(self, lora_id: int) -> bool:
247
248
def execute_dummy_batch(self) -> None:
249
runner = self.model_runner
250
- assert runner.dp_size > 1, "Dummy batch execution should only be " \
251
- "performed with data parallelism enabled, but got " \
252
- f"dp_size={runner.dp_size}."
+ if runner.dp_size <= 1:
+ raise ValueError("Dummy batch execution should only be "
+ "performed with data parallelism enabled, but got "
253
+ f"dp_size={runner.dp_size}.")
254
255
# If torchair graph is enabled, notify the other DP ranks that this is a
256
# dummy run by using '-1' as a flag for num_tokens. This will be
0 commit comments