Skip to content

Commit 0f2d197

Browse files
committed
init
Signed-off-by: Sage Moore <sage@neuralmagic.com>
1 parent 19d91ec commit 0f2d197

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

vllm/model_executor/layers/fused_moe/shared_fused_moe.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,13 +28,16 @@ def __init__(
2828
super().__init__(**kwargs)
2929
self._shared_experts = shared_experts
3030

31-
# Disable shared expert overlap if we are not using
31+
# Disable shared expert overlap if we are using eplb or not using
3232
# flashinfer + DP since there is nothing to be gained in this case.
3333
# Disabling the overlap optimization also prevents the shared experts
3434
# from being hidden from torch.compile.
3535
self.use_overlapped = (
3636
use_overlapped
37-
and not (self.use_flashinfer_cutlass_kernels and self.dp_size > 1)
37+
and not (
38+
self.enable_eplb
39+
or (self.use_flashinfer_cutlass_kernels and self.dp_size > 1)
40+
)
3841
and self._shared_experts is not None
3942
)
4043

0 commit comments

Comments
 (0)