Skip to content

Conversation

@wangxiyuan
Copy link
Collaborator

@wangxiyuan wangxiyuan commented Dec 1, 2025

  1. fix Update rope_scaling to rope_parameters in preparation for Transformers v5 vllm#28542
    The model structure modifications we involved in are:

    • Qwen2.5-VL(still exist some patch)
    • Qwen2-VL
    • Qwen2
    • DeepSeek series
    • Qwen-moe series
  2. fix Revert "[Redo] #26368 (#28771)" vllm#29121
    the output token now type changed from np to list[list[int]]

  3. fix [Core] Deprecate xformers vllm#29262
    xformers backend for multimodal now has been deprecated

  4. fix [Attention] Remove imports from vllm/attention/__init__.py vllm#29342

  5. fix [Core] Refactor padding logic and pad for CUDA graphs before attention metadata building  vllm#28579

  6. fix [Feature] Prefill Context Parallel (PCP) basic support vllm#28718

  7. fix [Config] Clean up SchedulerConfig initialization vllm#28665

  8. fix [Frontend][torch.compile] CompilationConfig Overhaul (#20283): Set up -O infrastructure vllm#26847
    vllm introduced the optimization-level, some default config has been changed, and the param --enforce-eager has been deprecated

  9. fix http://github.com/vllm-project/vllm/pull/29223 it retuns tuple for sampler.

  10. fix Remove upstream fa checks vllm#29471 we'll remove the related patch to avoid this kind of error.

Co-authored-by: hfadzxy starmoon_zhang@163.com
Co-authored-by: wangli wangli858794774@gmail.com

@github-actions
Copy link

github-actions bot commented Dec 1, 2025

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request upgrades vLLM to main, which involves a lot of refactoring to align with upstream changes. Most changes are related to module path updates, API signature changes (e.g., rope_parameters), and data type changes (from numpy arrays to Python lists). I've found a critical issue in vllm_ascend/spec_decode/eagle_proposer.py where a hardcoded index is used instead of iterating through the batch, which will lead to incorrect behavior in speculative decoding. Please address this issue.

@github-actions github-actions bot added the documentation Improvements or additions to documentation label Dec 1, 2025
@wangxiyuan wangxiyuan added ready read for review ready-for-test start test by label for PR labels Dec 1, 2025
@github-actions
Copy link

github-actions bot commented Dec 1, 2025

This pull request has conflicts, please resolve those before we can evaluate the pull request.

self.mock_vllm_config.scheduler_config = SchedulerConfig(
max_num_seqs=8, chunked_prefill_enabled=True)
mock_scheduler_config = MagicMock(spec=SchedulerConfig)
mock_scheduler_config.max_num_seqs = 8 # 设置为整数,不是 MagicMock
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

plz use english comment

Copy link
Collaborator

@MengqingCao MengqingCao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Plz fix the above comment and LGTM if CI passes

wangxiyuan and others added 12 commits December 2, 2025 17:03
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: hfadzxy <starmoon_zhang@163.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
wangxiyuan and others added 13 commits December 2, 2025 17:03
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
@wangxiyuan wangxiyuan merged commit 7f2673e into vllm-project:main Dec 2, 2025
24 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation module:core module:ops module:tests ready read for review ready-for-test start test by label for PR vllm-break

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants