Skip to content

Commit 7339042

Browse files
ganyi1996ppodevpatelio
authored andcommitted
[ROCm][BugFix] Fix shared expert loading error when disable VLLM_ROCM_USE_AITER_FUSION_SHARED_EXPERTS (vllm-project#28633)
Signed-off-by: ganyi <ygan@amd.com>
1 parent 83ce136 commit 7339042

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

vllm/model_executor/models/deepseek_v2.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -287,7 +287,10 @@ def __init__(
287287
)
288288

289289
self.is_rocm_aiter_moe_enabled = rocm_aiter_ops.is_fused_moe_enabled()
290-
if config.n_shared_experts is None or self.is_rocm_aiter_moe_enabled:
290+
self.is_fusion_moe_shared_experts_enabled = (
291+
rocm_aiter_ops.is_fusion_moe_shared_experts_enabled()
292+
)
293+
if config.n_shared_experts is None or self.is_fusion_moe_shared_experts_enabled:
291294
self.shared_experts = None
292295
else:
293296
intermediate_size = config.moe_intermediate_size * config.n_shared_experts
@@ -327,7 +330,7 @@ def __init__(
327330
num_redundant_experts=self.n_redundant_experts,
328331
is_sequence_parallel=self.is_sequence_parallel,
329332
n_shared_experts=config.n_shared_experts
330-
if rocm_aiter_ops.is_fusion_moe_shared_experts_enabled()
333+
if self.is_fusion_moe_shared_experts_enabled
331334
else None,
332335
)
333336

0 commit comments

Comments
 (0)