Skip to content

Commit 7f2673e

Browse files
wangxiyuanPotabkzhangxinyuehfad
authored
upgrade vLLM to main (#4608)
1. fix vllm-project/vllm#28542 The model structure modifications we involved in are: - Qwen2.5-VL(still exist some patch) - Qwen2-VL - Qwen2 - DeepSeek series - Qwen-moe series 2. fix vllm-project/vllm#29121 the output token now type changed from np to `list[list[int]]` 3. fix vllm-project/vllm#29262 `xformers` backend for multimodal now has been deprecated 4. fix vllm-project/vllm#29342 5. fix vllm-project/vllm#28579 6. fix vllm-project/vllm#28718 7. fix vllm-project/vllm#28665 8. fix vllm-project/vllm#26847 vllm introduced the `optimization-level`, some default config has been changed, and the param `--enforce-eager` has been deprecated 9. fix http://github.com/vllm-project/vllm/pull/29223 it retuns tuple for sampler. 10. fix vllm-project/vllm#29471 we'll remove the related patch to avoid this kind of error. Co-authored-by: hfadzxy <starmoon_zhang@163.com> Co-authored-by: wangli <wangli858794774@gmail.com> - vLLM version: v0.11.2 --------- Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com> Signed-off-by: wangli <wangli858794774@gmail.com> Signed-off-by: hfadzxy <starmoon_zhang@163.com> Co-authored-by: wangli <wangli858794774@gmail.com> Co-authored-by: hfadzxy <starmoon_zhang@163.com>
1 parent 4588cda commit 7f2673e

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

60 files changed

+383
-374
lines changed

.github/workflows/_e2e_nightly_multi_node.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ on:
3232
description: how many pods will be pulled up via lws.yaml, indicates number of nodes we need
3333
vllm_version:
3434
required: false
35-
default: "v0.11.2"
35+
default: "86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24"
3636
type: string
3737
description: vllm version to use
3838
vllm_ascend_remote_url:

.github/workflows/format_pr_body.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ jobs:
3636

3737
- name: Get vLLM version
3838
run: |
39-
VLLM_COMMIT=v0.11.2
39+
VLLM_COMMIT=86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
4040
echo "VLLM_COMMIT=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
4141
4242
- name: Checkout repository

.github/workflows/nightly_benchmarks.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ jobs:
5151
strategy:
5252
matrix:
5353
include:
54-
- vllm_branch: v0.11.2
54+
- vllm_branch: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
5555
vllm_ascend_branch: main
5656
max-parallel: 1
5757
container:

.github/workflows/vllm_ascend_test_nightly_a2.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ jobs:
8686
tests: tests/e2e/nightly/ops
8787
uses: ./.github/workflows/_e2e_nightly_single_node.yaml
8888
with:
89-
vllm: v0.11.2
89+
vllm: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
9090
runner: ${{ matrix.test_config.os }}
9191
tests: ${{ matrix.test_config.tests }}
9292
image: 'swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend:nightly-a2'
@@ -134,7 +134,7 @@ jobs:
134134
- Qwen3-Next-80B-A3B-Instruct
135135
uses: ./.github/workflows/_e2e_nightly_single_node_models.yaml
136136
with:
137-
vllm: v0.11.2
137+
vllm: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
138138
runner: ${{ matrix.test_config.os }}
139139
model_list: ${{ toJson(matrix.test_config.model_list) }}
140140
image: 'swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.3.rc2-910b-ubuntu22.04-py3.11'

.github/workflows/vllm_ascend_test_nightly_a3.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -139,7 +139,7 @@ jobs:
139139
tests: tests/e2e/nightly/models/test_glm4_5.py
140140
uses: ./.github/workflows/_e2e_nightly_single_node.yaml
141141
with:
142-
vllm: v0.11.2
142+
vllm: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
143143
runner: ${{ matrix.test_config.os }}
144144
image: 'swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend:nightly-a3'
145145
tests: ${{ matrix.test_config.tests }}

.github/workflows/vllm_ascend_test_pr_full.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ jobs:
6969
name: e2e-full
7070
strategy:
7171
matrix:
72-
vllm_version: [v0.11.2]
72+
vllm_version: [86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24]
7373
needs: [changes]
7474
if: ${{ needs.changes.outputs.e2e_tracker == 'true' }}
7575
uses: ./.github/workflows/_e2e_test.yaml

.github/workflows/vllm_ascend_test_pr_light.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ jobs:
4242
lint:
4343
uses: ./.github/workflows/pre-commit.yml
4444
with:
45-
vllm: v0.11.2
45+
vllm: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
4646
changes:
4747
runs-on: ubuntu-latest
4848
outputs:
@@ -84,7 +84,7 @@ jobs:
8484
SOC_VERSION: ascend910b1
8585
strategy:
8686
matrix:
87-
vllm_version: [v0.11.2]
87+
vllm_version: [86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24]
8888
steps:
8989
- name: Install packages
9090
run: |
@@ -142,7 +142,7 @@ jobs:
142142
name: e2e-light
143143
strategy:
144144
matrix:
145-
vllm_version: [v0.11.2]
145+
vllm_version: [86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24]
146146
# Note (yikun): If CI resource are limited we can split job into two chain jobs
147147
needs: [lint, changes]
148148
# only trigger e2e test after lint passed and the change is e2e related with pull request.

.github/workflows/vllm_ascend_test_report.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ jobs:
7272
- DeepSeek-V2-Lite
7373
uses: ./.github/workflows/_e2e_nightly_single_node_models.yaml
7474
with:
75-
vllm: v0.11.2
75+
vllm: 86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
7676
runner: ${{ matrix.runner }}
7777
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.3.rc2-910b-ubuntu22.04-py3.11
7878
model_list: ${{ toJson(matrix.model_list) }}

Dockerfile

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,8 +48,10 @@ RUN pip config set global.index-url ${PIP_INDEX_URL}
4848

4949
# Install vLLM
5050
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
51-
ARG VLLM_TAG=v0.11.2
52-
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
51+
ARG VLLM_TAG=86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
52+
# Revert this change once VLLM_TAG is specified to branch or tag
53+
# RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
54+
RUN git clone $VLLM_REPO /vllm-workspace/vllm && (cd /vllm-workspace/vllm && git checkout $VLLM_TAG)
5355
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.
5456
RUN VLLM_TARGET_DEVICE="empty" python3 -m pip install -v -e /vllm-workspace/vllm/[audio] --extra-index https://download.pytorch.org/whl/cpu/ && \
5557
python3 -m pip uninstall -y triton && \

Dockerfile.310p

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,10 @@ RUN pip config set global.index-url ${PIP_INDEX_URL}
3939

4040
# Install vLLM
4141
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
42-
ARG VLLM_TAG=v0.11.2
43-
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
42+
ARG VLLM_TAG=86e178f7c4d8c3b0eaf3c8e3f810a83f63b90e24
43+
# Revert this change once VLLM_TAG is specified to branch or tag
44+
# RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
45+
RUN git clone $VLLM_REPO /vllm-workspace/vllm && (cd /vllm-workspace/vllm && git checkout $VLLM_TAG)
4446
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.
4547
RUN VLLM_TARGET_DEVICE="empty" python3 -m pip install -v -e /vllm-workspace/vllm/[audio] --extra-index https://download.pytorch.org/whl/cpu/ && \
4648
python3 -m pip uninstall -y triton && \

0 commit comments

Comments
 (0)