Skip to content

Commit fdc4236

Browse files
committed
fix pp for llama4
Signed-off-by: Lu Fang <[email protected]>
1 parent 54a66e5 commit fdc4236

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/model_executor/models/mllama4.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -672,9 +672,9 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
672672
self.config,
673673
None,
674674
prefix=maybe_prefix(prefix, "multi_modal_projector"))
675-
676675
self.language_model = _initialize_model(
677-
vllm_config=vllm_config.with_hf_config(config.text_config),
676+
vllm_config=vllm_config.with_hf_config(config.text_config,
677+
["LlamaForCausalLM"]),
678678
prefix=maybe_prefix(prefix, "language_model"),
679679
model_class=Llama4ForCausalLM,
680680
)
@@ -824,7 +824,7 @@ def load_weights(self, weights: Iterable[Tuple[str,
824824
# language_model is an Llama4ForCausalLM instance. We load it's
825825
# using llama4's load_weights routine.
826826
language_model_weights, other_weights = self.separate_weights(
827-
weights, prefix="language_model.model.")
827+
weights, prefix="language_model.")
828828
loader = AutoWeightsLoader(self)
829829
loaded_language_model_params = loader.load_weights(
830830
language_model_weights)

0 commit comments

Comments
 (0)