@@ -117,16 +117,14 @@ steps:
117
117
commands :
118
118
- pytest -v -s core
119
119
120
- - label : Entrypoints Test # 40min
120
+ - label : Entrypoints Test (LLM) # 40min
121
121
mirror_hardwares : [amdexperimental]
122
122
working_dir : " /vllm-workspace/tests"
123
123
fast_check : true
124
124
torch_nightly : true
125
125
source_file_dependencies :
126
126
- vllm/
127
127
- tests/entrypoints/llm
128
- - tests/entrypoints/openai
129
- - tests/entrypoints/test_chat_utils
130
128
- tests/entrypoints/offline_mode
131
129
commands :
132
130
- export VLLM_WORKER_MULTIPROC_METHOD=spawn
@@ -135,9 +133,21 @@ steps:
135
133
- pytest -v -s entrypoints/llm/test_generate.py # it needs a clean process
136
134
- pytest -v -s entrypoints/llm/test_generate_multiple_loras.py # it needs a clean process
137
135
- VLLM_USE_V1=0 pytest -v -s entrypoints/llm/test_guided_generate.py # it needs a clean process
136
+ - VLLM_USE_V1=0 pytest -v -s entrypoints/offline_mode # Needs to avoid interference with other tests
137
+
138
+ - label : Entrypoints Test (API Server) # 40min
139
+ mirror_hardwares : [amdexperimental]
140
+ working_dir : " /vllm-workspace/tests"
141
+ fast_check : true
142
+ torch_nightly : true
143
+ source_file_dependencies :
144
+ - vllm/
145
+ - tests/entrypoints/openai
146
+ - tests/entrypoints/test_chat_utils
147
+ commands :
148
+ - export VLLM_WORKER_MULTIPROC_METHOD=spawn
138
149
- pytest -v -s entrypoints/openai --ignore=entrypoints/openai/test_chat_with_tool_reasoning.py --ignore=entrypoints/openai/test_oot_registration.py --ignore=entrypoints/openai/test_tensorizer_entrypoint.py --ignore=entrypoints/openai/correctness/
139
150
- pytest -v -s entrypoints/test_chat_utils.py
140
- - VLLM_USE_V1=0 pytest -v -s entrypoints/offline_mode # Needs to avoid interference with other tests
141
151
142
152
- label : Distributed Tests (4 GPUs) # 10min
143
153
mirror_hardwares : [amdexperimental]
0 commit comments