Skip to content

Commit ac63832

Browse files
MiterV1linhuang
authored andcommitted
[Bugfix] Number of shared_experts_fusion should set by user.
Number of shared_experts_fusion should set by user. Not enable DeepSeek V3 shared_experts_fusion in sm90 (sgl-project#5571) by default. Signed-off-by: miter <[email protected]>
1 parent 2c3b71d commit ac63832

File tree

1 file changed

+1
-3
lines changed

1 file changed

+1
-3
lines changed

python/sglang/srt/models/deepseek_v2.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1690,11 +1690,9 @@ def determine_n_share_experts_fusion(
16901690
and self.config.n_routed_experts == 256
16911691
and (not global_server_args_dict["enable_deepep_moe"])
16921692
):
1693-
self.n_share_experts_fusion = self.tp_size
1694-
global_server_args_dict["n_share_experts_fusion"] = self.tp_size
16951693
log_info_on_rank0(
16961694
logger,
1697-
"Deepseek V3/R1 with fp8 can use shared experts fusion optimization when SM version >=90. Shared experts fusion optimization is enabled.",
1695+
"Deepseek V3/R1 with fp8 can use shared experts fusion optimization when SM version >=90. Shared experts fusion optimization is not enabled.",
16981696
)
16991697

17001698
def get_input_embeddings(self) -> nn.Embedding:

0 commit comments

Comments
 (0)