Skip to content

Conversation

@biswapanda
Copy link
Contributor

@biswapanda biswapanda commented Dec 2, 2025

Overview:

Details:

Where should the reviewer start?

Related Issues: (use one of the action keywords Closes / Fixes / Resolves / Relates to)

  • closes GitHub issue: #xxx

Summary by CodeRabbit

  • Tests
    • Updated test infrastructure configuration.

Note: This release contains no user-facing changes. The modification is internal testing infrastructure only.

✏️ Tip: You can customize this high-level summary in your review settings.

@biswapanda biswapanda requested review from a team as code owners December 2, 2025 04:26
@github-actions github-actions bot added the fix label Dec 2, 2025
@biswapanda biswapanda changed the title fix: WAR to suppress test_serve_deployment for main CI fix fix: [ci] WAR to suppress test_serve_deployment in main Dec 2, 2025
@coderabbitai
Copy link
Contributor

coderabbitai bot commented Dec 2, 2025

Walkthrough

The @pytest.mark.gpu_1 decorator is commented out on the test_serve_deployment function in the test suite with an explanatory note, effectively removing the GPU-1 test requirement while preserving all other markers and test functionality.

Changes

Cohort / File(s) Summary
Test Marker Removal
tests/serve/test_vllm.py
Commented out @pytest.mark.gpu_1 on test_serve_deployment with a clarifying note; other test markers and logic remain intact

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~3 minutes

Poem

A marker removed, a constraint set free,
One line commented out, as simple can be!
No logic twisted, no function askew,
Just GPU-1 gone—a test that's made new! 🐰✨

Pre-merge checks

❌ Failed checks (1 warning)
Check name Status Explanation Resolution
Description check ⚠️ Warning The pull request description contains only template placeholders with no actual content filled in; all sections are empty or contain placeholder text like '#xxx'. Fill in the Overview, Details, and Where should the reviewer start sections with actual information about the workaround and its rationale. Replace placeholder '#xxx' with the actual related GitHub issue number.
✅ Passed checks (2 passed)
Check name Status Explanation
Title check ✅ Passed The title clearly describes the main change: disabling a GPU test marker (WAR = workaround) for a specific test to suppress it in CI.
Docstring Coverage ✅ Passed Docstring coverage is 100.00% which is sufficient. The required threshold is 80.00%.

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

📜 Review details

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 5708b70 and 4e87250.

📒 Files selected for processing (1)
  • tests/serve/test_vllm.py (1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (7)
  • GitHub Check: operator (amd64)
  • GitHub Check: sglang (amd64)
  • GitHub Check: operator (arm64)
  • GitHub Check: vllm (arm64)
  • GitHub Check: trtllm (amd64)
  • GitHub Check: vllm (amd64)
  • GitHub Check: Build and Test - dynamo

@pytest.mark.vllm
@pytest.mark.e2e
@pytest.mark.gpu_1
# @pytest.mark.gpu_1 # Disabled for now to unblock CI in main.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

🧩 Analysis chain

🏁 Script executed:

# Find params_with_model_mark function definition
rg -n "def params_with_model_mark" -A 15

Repository: ai-dynamo/dynamo

Length of output: 793


🏁 Script executed:

# Also search for where params_with_model_mark is imported or used
rg -n "params_with_model_mark" -B 2 -A 2

Repository: ai-dynamo/dynamo

Length of output: 2051


🏁 Script executed:

# Check for any issue references or TODOs in the file
rg -n "(TODO|FIXME|issue|#[0-9])" tests/serve/test_vllm.py

Repository: ai-dynamo/dynamo

Length of output: 296


🏁 Script executed:

# Look at the file structure to understand mark application
head -350 tests/serve/test_vllm.py | tail -50

Repository: ai-dynamo/dynamo

Length of output: 1833


This workaround is ineffective—config-level gpu_1 marks will still apply.

The params_with_model_mark() function (tests/serve/common.py:70) preserves all config-level marks by copying them into the parameterized test parameters (line 77). Commenting out the function-level @pytest.mark.gpu_1 does not disable GPU requirements for test instances whose configs explicitly include pytest.mark.gpu_1 in their marks array—specifically "aggregated" (line 46), "aggregated_lmcache" (line 58), "agg-request-plane-tcp" (line 71), and "agg-request-plane-http" (line 83).

Additionally, this temporary workaround lacks a tracking issue. The PR description should reference an issue for:

  1. Re-enabling the marker once CI is fixed
  2. Addressing the root cause of the GPU requirement conflict
🤖 Prompt for AI Agents
In tests/serve/test_vllm.py around line 340, commenting out the function-level
@pytest.mark.gpu_1 is ineffective because params_with_model_mark in
tests/serve/common.py preserves config-level marks (so configs like
"aggregated", "aggregated_lmcache", "agg-request-plane-tcp", and
"agg-request-plane-http" still force GPU). Fix by either removing or
conditionally filtering out pytest.mark.gpu_1 from the per-config marks in
params_with_model_mark (so a test-level disable actually takes effect), or add a
top-level check in test_vllm.py to skip parameterized cases whose combined marks
include gpu_1; also add a tracking issue reference in the PR description for (1)
re-enabling the marker when CI is fixed and (2) investigating/fixing the root
cause of the GPU requirement conflict.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants