Skip to content

Commit 566ec04

Browse files
Adding "Basic Models Test" and "Multi-Modal Models Test (Extended) 3" in AMD Pipeline (vllm-project#18106)
Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com> Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
1 parent 01c2233 commit 566ec04

File tree

4 files changed

+25
-3
lines changed

4 files changed

+25
-3
lines changed

.buildkite/scripts/hardware_ci/run-amd-test.sh

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -82,6 +82,14 @@ if [[ $commands == *"pytest -v -s basic_correctness/test_basic_correctness.py"*
8282
commands=${commands//"pytest -v -s basic_correctness/test_basic_correctness.py"/"VLLM_USE_TRITON_FLASH_ATTN=0 pytest -v -s basic_correctness/test_basic_correctness.py"}
8383
fi
8484

85+
if [[ $commands == *"pytest -v -s models/test_registry.py"* ]]; then
86+
commands=${commands//"pytest -v -s models/test_registry.py"/"pytest -v -s models/test_registry.py -k 'not BambaForCausalLM and not GritLM and not Mamba2ForCausalLM and not Zamba2ForCausalLM'"}
87+
fi
88+
89+
if [[ $commands == *"VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2'"* ]]; then
90+
commands=${commands//"VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2'"/"VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2 and not BambaForCausalLM and not Gemma2ForCausalLM and not Grok1ModelForCausalLM and not Zamba2ForCausalLM and not Gemma2Model and not GritLM'"}
91+
fi
92+
8593
if [[ $commands == *"pytest -v -s compile/test_basic_correctness.py"* ]]; then
8694
commands=${commands//"pytest -v -s compile/test_basic_correctness.py"/"VLLM_USE_TRITON_FLASH_ATTN=0 pytest -v -s compile/test_basic_correctness.py"}
8795
fi

.buildkite/test-pipeline.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -455,7 +455,7 @@ steps:
455455
##### models test #####
456456

457457
- label: Basic Models Test # 24min
458-
mirror_hardwares: [amdexperimental]
458+
mirror_hardwares: [amdexperimental, amdproduction]
459459
torch_nightly: true
460460
source_file_dependencies:
461461
- vllm/
@@ -527,7 +527,7 @@ steps:
527527
- pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=0) and not core_model'
528528

529529
- label: Multi-Modal Models Test (Extended) 3
530-
mirror_hardwares: [amdexperimental]
530+
mirror_hardwares: [amdexperimental, amdproduction]
531531
optional: true
532532
source_file_dependencies:
533533
- vllm/
@@ -537,7 +537,7 @@ steps:
537537
- pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=1) and not core_model'
538538

539539
- label: Quantized Models Test
540-
mirror_hardwares: [amdexperimental]
540+
mirror_hardwares: [amdexperimental, amdproduction]
541541
source_file_dependencies:
542542
- vllm/model_executor/layers/quantization
543543
- tests/models/quantization

requirements/rocm-test.txt

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,4 +22,10 @@ decord==0.6.0
2222
#sentence-transformers # required by entrypoints/openai/test_score.py
2323
sentence-transformers==3.4.1
2424

25+
# Basic Models Test
26+
matplotlib==3.10.3
27+
28+
# Multi-Modal Models Test (Extended) 3
29+
blobfile==3.0.0
30+
2531

tests/models/test_transformers.py

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,8 @@
22
"""Test the functionality of the Transformers backend."""
33
import pytest
44

5+
from vllm.platforms import current_platform
6+
57
from ..conftest import HfRunner, VllmRunner
68
from ..utils import multi_gpu_test
79
from .utils import check_logprobs_close
@@ -33,6 +35,9 @@ def check_implementation(
3335
)
3436

3537

38+
@pytest.mark.skipif(
39+
current_platform.is_rocm(),
40+
reason="Llama-3.2-1B-Instruct, Ilama-3.2-1B produce memory access fault.")
3641
@pytest.mark.parametrize(
3742
"model,model_impl",
3843
[
@@ -64,6 +69,9 @@ def test_distributed(
6469
"meta-llama/Llama-3.2-1B-Instruct", **kwargs)
6570

6671

72+
@pytest.mark.skipif(
73+
current_platform.is_rocm(),
74+
reason="bitsandbytes quantization is currently not supported in rocm.")
6775
@pytest.mark.parametrize("model, quantization_kwargs", [
6876
(
6977
"meta-llama/Llama-3.2-1B-Instruct",

0 commit comments

Comments
 (0)