diff --git a/vllm/model_executor/models/llava.py b/vllm/model_executor/models/llava.py index 65c6bd07bfff0..53eef72dd5f91 100644 --- a/vllm/model_executor/models/llava.py +++ b/vllm/model_executor/models/llava.py @@ -218,12 +218,8 @@ def _get_dummy_mm_kwargs( image_processor = hf_processor.image_processor # type: ignore hf_inputs = image_processor.preprocess(data['image'], return_tensors="pt") - is_pixtral = isinstance(hf_processor, PixtralProcessor) - return MultiModalKwargs( - **hf_inputs, - is_pixtral=torch.tensor(is_pixtral), - ) + return MultiModalKwargs(**hf_inputs) class LlavaLikeConfig(Protocol):