Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions .github/configs/nvidia-master.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3076,7 +3076,7 @@ gptoss-fp4-b200-trt:
- { tp: 8, conc-start: 4, conc-end: 4}

gptoss-fp4-b200-vllm:
image: vllm/vllm-openai:v0.15.1
image: vllm/vllm-openai:v0.16.0-cu130
model: openai/gpt-oss-120b
model-prefix: gptoss
runner: b200
Expand Down Expand Up @@ -3107,7 +3107,7 @@ gptoss-fp4-b200-vllm:
- { tp: 8, conc-start: 4, conc-end: 4 }

gptoss-fp4-h100-vllm:
image: vllm/vllm-openai:v0.15.1
image: vllm/vllm-openai:v0.16.0-cu130
model: openai/gpt-oss-120b
model-prefix: gptoss
runner: h100
Expand Down Expand Up @@ -3386,7 +3386,7 @@ gptoss-fp4-h200-trt:
- { tp: 8, ep: 8, dp-attn: false, conc-start: 4, conc-end: 8 }

gptoss-fp4-h200-vllm:
image: vllm/vllm-openai:v0.15.1
image: vllm/vllm-openai:v0.16.0-cu130
model: openai/gpt-oss-120b
model-prefix: gptoss
runner: h200
Expand Down
7 changes: 7 additions & 0 deletions perf-changelog.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -932,3 +932,10 @@
- "Remove deprecated VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION/VLLM_ROCM_USE_AITER_MHA env vars and compilation-config cudagraph_mode"
pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/867

- config-keys:
- gptoss-fp4-b200-vllm
- gptoss-fp4-h100-vllm
- gptoss-fp4-h200-vllm
description:
- "Update vLLM image from v0.15.1 to v0.16.0-cu130 for NVIDIA GPT-OSS configs"
pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/800
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🟡 The pr-link for the new changelog entry points to PR #800, which was already merged and then reverted (PR #903). Since this PR #904 is the one actually re-landing the change, the link should be #904 for proper traceability.

Extended reasoning...

What the bug is

The new perf-changelog.yaml entry added at the bottom of the file has its pr-link set to #800. However, PR #800 was previously merged (commit da55158), then reverted by PR #903 (commit cad1169). This PR #904 is the one that actually re-lands the vLLM image update (now with the -cu130 suffix), so the changelog entry should reference #904.

Step-by-step proof

  1. PR [NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0 #800 was merged as commit da55158, updating the NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0.
  2. PR Revert "[NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0" #903 reverted that change (commit cad1169: Revert [NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0 ([NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0 #800)).
  3. The revert was then reverted (commit c28ce20), and this PR [NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0 #904 re-lands the change with a slightly different image tag (v0.16.0-cu130 instead of v0.16.0).
  4. The diff shows line 941: pr-link: [NVIDIA] Update NVIDIA GPT-OSS vLLM image from v0.15.1 to v0.16.0 #800 — this references the reverted PR.

Why existing code does not prevent it

There is no automated validation that pr-link values reference the current PR. The author likely copied the link from the original PR #800 submission without updating it for the re-landing PR.

Impact

Anyone following the changelog link to understand this change would land on PR #800, which is marked as reverted. This is confusing for traceability, though it does not affect any functional behavior. The convention throughout perf-changelog.yaml is that pr-link references the PR that actually lands the change.

Fix

Change line 941 from:

pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/800

to:

pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/904

Loading