We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3c259e4 commit da80873Copy full SHA for da80873
vllm-benchmarks/run.sh
@@ -41,9 +41,8 @@ build_vllm() {
41
SCCACHE_CACHE_SIZE=100G sccache --start-server || true
42
# Build and install vLLM
43
if command -v nvidia-smi; then
44
- pip install -r requirements/build.txt
45
- # TODO (huydhn): https://github.com/vllm-project/vllm/issues/15435
46
- TORCH_CUDA_ARCH_LIST=8.0 pip install --editable .
+ pip install -r requirements/build.txt --extra-index-url https://download.pytorch.org/whl/cu128
+ pip install --editable .
47
elif command -v amd-smi; then
48
pip install -r requirements/rocm.txt
49
pip install -r requirements/rocm-build.txt
0 commit comments