Skip to content

[NOT FOR LANDING] 355_wip_0909_rc2 -> 0909_rc2#654

Draft
maleksan85 wants to merge 974 commits into0909_rc2from
355_wip_0909_rc2
Draft

[NOT FOR LANDING] 355_wip_0909_rc2 -> 0909_rc2#654
maleksan85 wants to merge 974 commits into0909_rc2from
355_wip_0909_rc2

Conversation

@maleksan85
Copy link

@maleksan85 maleksan85 commented Sep 4, 2025

Please direct your PRs to the upstream vllm (https://github.com/vllm-project/vllm.git)

Accepting PRs into the ROCm fork (https://github.com/ROCm/vllm) will require a clear previously communicated exception

vllmellm and others added 30 commits April 3, 2025 04:45
Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
* Adding 2stage MoE support separately until it is added upstream

* Missing layour param
* Enable fused fp8 out in V1 CPA and FA

* Correct operation and creating the tensot or th correct type

* Update to use for the non-custom path as well

* This was a debug assert
* Added the extra use_irope parameter in

Co-authored-by: Hongxia Yang <hongxia.yang@amd.com>
Signed-off-by: tjtanaa <tunjian.tan@embeddedllm.com>

* Fix ROCm V1 Engine Fused MoE Bug

Signed-off-by: tjtanaa <tunjian.tan@embeddedllm.com>

* Add warning message that V0 do not support irope

Signed-off-by: tjtanaa <tunjian.tan@embeddedllm.com>

---------

Signed-off-by: tjtanaa <tunjian.tan@embeddedllm.com>
Co-authored-by: tjtanaa <tunjian.tan@embeddedllm.com>
Co-authored-by: Hongxia Yang <hongxia.yang@amd.com>
add RAY_EXPERIMENTAL_NOSET_ROCR_VISIBLE_DEVICES=1
Signed-off-by: maleksan85 <maleksan@amd.com>
Co-authored-by: maleksan85 <maleksan@amd.com>
Signed-off-by: charlifu <charlifu@amd.com>
Signed-off-by: jpvillam <Juan.Villamizar@amd.com>
Signed-off-by: jpvillam <Juan.Villamizar@amd.com>
* Updated README.md with April 10 results

* Updated README.md with "2-stage MoE and MLA from AITER"
Added correct path for Dockerfile.rocm under Docker manifest
Signed-off-by: charlifu <charlifu@amd.com>
The upstream has moved docker files into a separate directory.

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>
* Update test-template.j2 to fix new location of run-amd-test.sh

Update test-template.j2 to fix new location of run-amd-test.sh

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

* Update test-template.j2 upsie, wrong path initially

Update test-template.j2 upsie, wrong path initially

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

* Update test-template.j2

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

---------

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>
* Adapted hipblaslt build to work with rocm 6.4

* rccl version compatible with 6.4

* Torch and triton combination that works

* hipblaslt version and not rebuilding rccl

* Fixing another package that we install now
Added Known Issues section to document meta 405B FP 8 model mem fault and work around.
Signed-off-by: seungrokjung <seungrok.jung@amd.com>
Mcirino1 and others added 16 commits September 1, 2025 05:15
* Updated README.md for June 10 release

* Added Docker Manifest git hash
* Updated README.md for June 24 Docker release

* Added additional throughput results

* Fixed some throughput results
* Minor changes to command line examples

* README changes and added throughput results

Still waiting on latency

* Added latency results

* Update README.md

* Update README.md
* Update test-pipeline.yaml

Disabling the "Tensorizer Test".

The test is seen to generate exceptions while still reporting as successful. That needs to be verified before re-enabling the test in the production environment.

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

* Fixing pre-commit complaints.

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

* .

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>

---------

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>
@maleksan85 maleksan85 changed the title 355 wip 0909 rc2 355_wip_0909_rc2 -> 0909_rc2 Sep 4, 2025
@maleksan85 maleksan85 changed the title 355_wip_0909_rc2 -> 0909_rc2 [NOT FOR LANDING] 355_wip_0909_rc2 -> 0909_rc2 Sep 4, 2025
attn_type: AttentionType = AttentionType.DECODER,
kv_sharing_target_layer_name: Optional[int] = None,
sinks: Optional[torch.Tensor] = None,
sinks: Optional[torch.Tensor] = None,
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fix it!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.