Skip to content

Commit c69ceac

Browse files
authored
[Doc][0.7.3] Add release note for 0.7.3.post1 (#1010)
cherry-pick 0.7.3.post1 release note from main Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
1 parent 0bfa79c commit c69ceac

File tree

4 files changed

+27
-6
lines changed

4 files changed

+27
-6
lines changed

docs/source/conf.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -67,10 +67,10 @@
6767
# the branch of vllm-ascend, used in vllm-ascend clone and image tag
6868
# - main branch: 'main'
6969
# - vX.Y.Z branch: latest vllm-ascend release tag
70-
'vllm_ascend_version': 'v0.7.3',
70+
'vllm_ascend_version': 'v0.7.3.post1',
7171
# the newest release version of vllm-ascend and matched vLLM, used in pip install.
7272
# This value should be updated when cut down release.
73-
'pip_vllm_ascend_version': "0.7.3",
73+
'pip_vllm_ascend_version': "0.7.3.post1",
7474
'pip_vllm_version': "0.7.3",
7575
# The maching MindIE Turbo for vLLM Ascend
7676
'pip_mindie_turbo_version': "2.0rc1",

docs/source/developer_guide/versioning_policy.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -67,6 +67,7 @@ Following is the Release Compatibility Matrix for vLLM Ascend Plugin:
6767

6868
| vllm-ascend | vLLM | Python | Stable CANN | PyTorch/torch_npu | MindIE Turbo |
6969
|--------------|--------------| --- | --- | --- | --- |
70+
| v0.7.3.post1 | v0.7.3 | 3.9 - 3.11 | 8.1.0 | 2.5.1 / 2.5.1 | 2.0rc1 |
7071
| v0.7.3 | v0.7.3 | 3.9 - 3.11 | 8.1.0 | 2.5.1 / 2.5.1 | 2.0rc1 |
7172
| v0.7.3rc2 | v0.7.3 | 3.9 - 3.11 | 8.0.0 | 2.5.1 / 2.5.1.dev20250320 | / |
7273
| v0.7.3rc1 | v0.7.3 | 3.9 - 3.11 | 8.0.0 | 2.5.1 / 2.5.1.dev20250308 | / |
@@ -78,6 +79,7 @@ Following is the Release Compatibility Matrix for vLLM Ascend Plugin:
7879

7980
| Date | Event |
8081
|------------|-------------------------------------------|
82+
| 2025.05.29 | Final release, v0.7.3.post1 |
8183
| 2025.05.08 | Final release, v0.7.3 |
8284
| 2025.04.17 | Release candidates, v0.8.4rc1 |
8385
| 2025.03.28 | Release candidates, v0.7.3rc2 |

docs/source/faqs.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2,10 +2,7 @@
22

33
## Version Specific FAQs
44

5-
- [[v0.7.1rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/19)
6-
- [[v0.7.3rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/267)
7-
- [[v0.7.3rc2] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/418)
8-
- [[v0.8.4rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/546)
5+
- [[v0.7.3.post1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/1007)
96

107
## General FAQs
118

docs/source/user_guide/release_notes.md

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,27 @@
11
# Release note
22

3+
## v0.7.3.post1
4+
5+
This is the first post release of 0.7.3. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev) to start the journey. It includes the following changes:
6+
7+
### Highlights
8+
9+
- Qwen3 and Qwen3MOE is supported now. The performance and accuracy of Qwen3 is well tested. You can try it now. Mindie Turbo is recomanded to improve the performance of Qwen3. [#903](https://github.com/vllm-project/vllm-ascend/pull/903) [#915](https://github.com/vllm-project/vllm-ascend/pull/915)
10+
- Added a new performance guide. The guide aims to help users to improve vllm-ascend performance on system level. It includes OS configuration, library optimization, deploy guide and so on. [#878](https://github.com/vllm-project/vllm-ascend/pull/878) [Doc Link](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev/developer_guide/performance/optimization_and_tuning.html)
11+
12+
### Bug Fix
13+
14+
- Qwen2.5-VL works for RLHF scenarios now. [#928](https://github.com/vllm-project/vllm-ascend/pull/928)
15+
- Users can launch the model from online weights now. e.g. from huggingface or modelscope directly [#858](https://github.com/vllm-project/vllm-ascend/pull/858) [#918](https://github.com/vllm-project/vllm-ascend/pull/918)
16+
- The meaningless log info `UserWorkspaceSize0` has been cleaned. [#911](https://github.com/vllm-project/vllm-ascend/pull/911)
17+
- The log level for `Failed to import vllm_ascend_C` has been changed to `warning` instead of `error`. [#956](https://github.com/vllm-project/vllm-ascend/pull/956)
18+
- DeepSeek MLA now works with chunked prefill in V1 Engine. Please note that V1 engine in 0.7.3 is just expermential and only for test usage. [#849](https://github.com/vllm-project/vllm-ascend/pull/849) [#936](https://github.com/vllm-project/vllm-ascend/pull/936)
19+
20+
### Docs
21+
22+
- The benchmark doc is updated for Qwen2.5 and Qwen2.5-VL [#792](https://github.com/vllm-project/vllm-ascend/pull/792)
23+
- Add the note to clear that only "modelscope<1.23.0" works with 0.7.3. [#954](https://github.com/vllm-project/vllm-ascend/pull/954)
24+
325
## v0.7.3
426

527
🎉 Hello, World!

0 commit comments

Comments
 (0)