Skip to content

Commit 56e58fe

Browse files
committed
fix ci
1 parent 74b96fa commit 56e58fe

File tree

2 files changed

+4
-1
lines changed

2 files changed

+4
-1
lines changed

python/sglang/srt/models/llama4.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646
from sglang.srt.layers.rotary_embedding import get_rope
4747
from sglang.srt.layers.vocab_parallel_embedding import VocabParallelEmbedding
4848
from sglang.srt.managers.schedule_batch import global_server_args_dict
49-
from sglang.srt.model_executor.forward_batch_info import ForwardBatch
49+
from sglang.srt.model_executor.forward_batch_info import ForwardBatch, PPProxyTensors
5050
from sglang.srt.models.llama import LlamaForCausalLM, LlamaMLP
5151
from sglang.srt.utils import add_prefix, fast_topk, get_compiler_backend, make_layers
5252

@@ -431,6 +431,7 @@ def forward(
431431
positions: torch.Tensor,
432432
forward_batch: ForwardBatch,
433433
input_embeds: torch.Tensor = None,
434+
pp_proxy_tensors: Optional[PPProxyTensors] = None,
434435
) -> Union[torch.Tensor, Tuple[torch.Tensor, List[torch.Tensor]]]:
435436
if input_embeds is None:
436437
hidden_states = self.embed_tokens(input_ids)

test/srt/test_vlm_accuracy.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -147,6 +147,8 @@ def get_sglang_model(self):
147147
gpu_id=0,
148148
tp_rank=0,
149149
tp_size=1,
150+
pp_rank=0,
151+
pp_size=1,
150152
nccl_port=12435,
151153
server_args=ServerArgs(
152154
model_path=self.model_path,

0 commit comments

Comments
 (0)