Skip to content

Commit 359200f

Browse files
authored
[doc] fix link (#20417)
Signed-off-by: reidliu41 <reid201711@gmail.com>
1 parent 220aee9 commit 359200f

File tree

1 file changed

+1
-1
lines changed
  • examples/offline_inference/profiling_tpu

1 file changed

+1
-1
lines changed

examples/offline_inference/profiling_tpu/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ This script is used to profile the TPU performance of vLLM for specific prefill
44

55
Note: an actual running server is a mix of both prefill of many shapes and decode of many shapes.
66

7-
We assume you are on a TPU already (this was tested on TPU v6e) and have installed vLLM according to the [installation guide](https://docs.vllm.ai/en/latest/getting_started/installation/ai_accelerator/index.html).
7+
We assume you are on a TPU already (this was tested on TPU v6e) and have installed vLLM according to the [Google TPU installation guide](https://docs.vllm.ai/en/latest/getting_started/installation/google_tpu.html).
88

99
> In all examples below, we run several warmups before (so `--enforce-eager` is okay)
1010

0 commit comments

Comments
 (0)