Skip to content

Commit a5d74ea

Browse files
committed
Merge branch 'smolvlm-for-test' into lkchen-ray_data_llm
Signed-off-by: Linkun Chen <github@lkchen.net>
2 parents 5a80423 + e4d60f2 commit a5d74ea

File tree

2 files changed

+26
-2
lines changed

2 files changed

+26
-2
lines changed

python/ray/llm/tests/batch/gpu/processor/test_vllm_engine_proc.py

+4-2
Original file line numberDiff line numberDiff line change
@@ -176,13 +176,15 @@ def test_embedding_model(gpu_type, model_opt_125m):
176176
assert all("prompt" in out for out in outs)
177177

178178

179-
def test_vision_model_v0(gpu_type, model_llava_354m):
179+
def test_vision_model(gpu_type, model_smolvlm_256m):
180180
processor_config = vLLMEngineProcessorConfig(
181-
model_source=model_llava_354m,
181+
model_source=model_smolvlm_256m,
182182
task_type="generate",
183183
engine_kwargs=dict(
184184
# Skip CUDA graph capturing to reduce startup time.
185185
enforce_eager=True,
186+
# # CI uses T4 GPU which does not support bfloat16.
187+
dtype="half",
186188
),
187189
# CI uses T4 GPU which is not supported by vLLM v1 FlashAttn.
188190
runtime_env=dict(

python/ray/llm/tests/conftest.py

+22
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@
55
from typing import Generator, List
66

77
S3_ARTIFACT_URL = "https://air-example-data.s3.amazonaws.com/"
8+
S3_ARTIFACT_LLM_OSSCI_URL = S3_ARTIFACT_URL + "rayllm-ossci/"
89

910

1011
def download_model_from_s3(
@@ -69,6 +70,27 @@ def model_llava_354m():
6970
yield from download_model_from_s3(REMOTE_URL, FILE_LIST)
7071

7172

73+
@pytest.fixture(scope="session")
74+
def model_smolvlm_256m():
75+
"""The vision language model for testing."""
76+
REMOTE_URL = f"{S3_ARTIFACT_LLM_OSSCI_URL}smolvlm-256m-instruct/"
77+
FILE_LIST = [
78+
"added_tokens.json",
79+
"chat_template.json",
80+
"config.json",
81+
"generation_config.json",
82+
"merges.txt",
83+
"model.safetensors",
84+
"preprocessor_config.json",
85+
"processor_config.json",
86+
"special_tokens_map.json",
87+
"tokenizer.json",
88+
"tokenizer_config.json",
89+
"vocab.json",
90+
]
91+
yield from download_model_from_s3(REMOTE_URL, FILE_LIST)
92+
93+
7294
@pytest.fixture(scope="session")
7395
def model_llama_3_2_216M():
7496
"""The llama 3.2 216M model for testing."""

0 commit comments

Comments
 (0)