@@ -41,16 +41,10 @@ concurrency:
41
41
42
42
jobs :
43
43
lint :
44
- # Only trigger lint on pull request
45
- if : ${{ github.event_name == 'pull_request' }}
46
44
uses : ./.github/workflows/pre-commit.yml
47
45
48
46
changes :
49
- # Only trigger changes on pull request
50
- if : ${{ github.event_name == 'pull_request' }}
51
47
runs-on : ubuntu-latest
52
- permissions :
53
- pull-requests : read
54
48
outputs :
55
49
e2e_tracker : ${{ steps.filter.outputs.e2e_tracker }}
56
50
ut_tracker : ${{ steps.filter.outputs.ut_tracker }}
@@ -60,20 +54,24 @@ jobs:
60
54
with :
61
55
filters : |
62
56
e2e_tracker:
57
+ - '.github/workflows/vllm_ascend_test.yaml'
63
58
- 'vllm_ascend/**'
64
59
- 'csrc/**'
65
60
- 'cmake/**'
66
61
- 'tests/e2e/**'
67
- - 'tests/conftest.py'
68
- - 'tests/model_utils.py'
69
- - 'tests/utils.py'
62
+ - 'CMakeLists.txt'
63
+ - 'setup.py'
64
+ - 'requirements.txt'
65
+ - 'requirements-dev.txt'
66
+ - 'requirements-lint.txt'
67
+ - 'packages.txt'
70
68
ut_tracker:
71
69
- 'tests/ut/**'
72
70
ut :
73
71
needs : [lint, changes]
74
72
name : unit test
75
- # only trigger unit test after lint passed and the change is e2e and ut related. Or the PR is merged.
76
- if : ${{ github.event_name == 'push' || ( needs.lint.result == 'success' && (needs.changes.outputs.e2e_tracker == 'true' || needs.changes.outputs.ut_tracker == 'true') ) }}
73
+ # only trigger unit test after lint passed and the change is e2e and ut related.
74
+ if : ${{ needs.lint.result == 'success' && (needs.changes.outputs.e2e_tracker == 'true' || needs.changes.outputs.ut_tracker == 'true') }}
77
75
runs-on : ubuntu-latest
78
76
container :
79
77
image : quay.io/ascend/cann:8.1.rc1-910b-ubuntu22.04-py3.10
@@ -112,9 +110,8 @@ jobs:
112
110
python3 -m pip install -r requirements-dev.txt --extra-index https://download.pytorch.org/whl/cpu/
113
111
python3 -m pip install -v . --extra-index https://download.pytorch.org/whl/cpu/
114
112
115
- - name : Run unit test for V1 Engine
113
+ - name : Run unit test
116
114
env :
117
- VLLM_USE_V1 : 1
118
115
VLLM_WORKER_MULTIPROC_METHOD : spawn
119
116
TORCH_DEVICE_BACKEND_AUTOLOAD : 0
120
117
run : |
@@ -133,8 +130,8 @@ jobs:
133
130
134
131
e2e :
135
132
needs : [lint, changes]
136
- # only trigger e2e test after lint passed and the change is e2e related.
137
- if : ${{ needs.lint.result == 'success' && needs.changes.outputs.e2e_tracker == 'true' }}
133
+ # only trigger e2e test after lint passed and the change is e2e related with pull request .
134
+ if : ${{ github.event_name == 'pull_request' && needs.lint.result == 'success' && needs.changes.outputs.e2e_tracker == 'true' }}
138
135
strategy :
139
136
max-parallel : 2
140
137
matrix :
@@ -189,9 +186,8 @@ jobs:
189
186
pip install -r requirements-dev.txt
190
187
pip install -v -e .
191
188
192
- - name : Run e2e test for V1 Engine
189
+ - name : Run e2e test
193
190
env :
194
- VLLM_USE_V1 : 1
195
191
VLLM_WORKER_MULTIPROC_METHOD : spawn
196
192
VLLM_USE_MODELSCOPE : True
197
193
run : |
@@ -213,26 +209,6 @@ jobs:
213
209
# TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
214
210
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
215
211
216
- - name : Run e2e test on V0 engine
217
- if : ${{ github.event_name == 'schedule' }}
218
- env :
219
- VLLM_USE_V1 : 0
220
- VLLM_USE_MODELSCOPE : True
221
- run : |
222
- pytest -sv tests/e2e/singlecard/test_offline_inference.py
223
- pytest -sv tests/e2e/singlecard/test_ilama_lora.py
224
- pytest -sv tests/e2e/singlecard/test_guided_decoding.py
225
- pytest -sv tests/e2e/singlecard/test_camem.py
226
- pytest -sv tests/e2e/singlecard/test_prompt_embedding.py
227
- pytest -sv tests/e2e/singlecard/test_embedding.py
228
- pytest -sv tests/e2e/singlecard/ \
229
- --ignore=tests/e2e/singlecard/test_offline_inference.py \
230
- --ignore=tests/e2e/singlecard/test_ilama_lora.py \
231
- --ignore=tests/e2e/singlecard/test_guided_decoding.py \
232
- --ignore=tests/e2e/singlecard/test_camem.py \
233
- --ignore=tests/e2e/singlecard/test_prompt_embedding.py \
234
- --ignore=tests/e2e/singlecard/test_embedding.py
235
-
236
212
e2e-4-cards :
237
213
needs : [e2e]
238
214
if : ${{ needs.e2e.result == 'success' }}
@@ -290,9 +266,8 @@ jobs:
290
266
pip install -r requirements-dev.txt
291
267
pip install -v -e .
292
268
293
- - name : Run vllm-project/vllm-ascend test for V1 Engine
269
+ - name : Run vllm-project/vllm-ascend test
294
270
env :
295
- VLLM_USE_V1 : 1
296
271
VLLM_WORKER_MULTIPROC_METHOD : spawn
297
272
VLLM_USE_MODELSCOPE : True
298
273
run : |
@@ -308,19 +283,3 @@ jobs:
308
283
pytest -sv tests/e2e/multicard/ --ignore=tests/e2e/multicard/test_ilama_lora_tp2.py \
309
284
--ignore=tests/e2e/multicard/test_offline_inference_distributed.py \
310
285
--ignore=tests/e2e/multicard/test_data_parallel.py
311
-
312
- - name : Run vllm-project/vllm-ascend test on V0 engine
313
- if : ${{ github.event_name == 'schedule' }}
314
- env :
315
- VLLM_USE_V1 : 0
316
- VLLM_USE_MODELSCOPE : True
317
- run : |
318
- pytest -sv tests/e2e/multicard/test_ilama_lora_tp2.py
319
- # Fixme: run VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py will raise error.
320
- # To avoid oom, we need to run the test in a single process.
321
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
322
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
323
- pytest -sv tests/e2e/multicard/test_data_parallel.py
324
- pytest -sv tests/e2e/multicard/ --ignore=tests/e2e/multicard/test_ilama_lora_tp2.py \
325
- --ignore=tests/e2e/multicard/test_offline_inference_distributed.py \
326
- --ignore=tests/e2e/multicard/test_data_parallel.py
0 commit comments