From 1e8558e26986fdf64581d7019db4ef1b8af79976 Mon Sep 17 00:00:00 2001 From: Flavia Beo Date: Fri, 25 Oct 2024 16:16:42 -0300 Subject: [PATCH] Fix for encode failing tests Signed-off-by: Flavia Beo --- vllm/model_executor/models/llama.py | 1 + 1 file changed, 1 insertion(+) diff --git a/vllm/model_executor/models/llama.py b/vllm/model_executor/models/llama.py index c346e3e808e3f..dfd75a60c67a6 100644 --- a/vllm/model_executor/models/llama.py +++ b/vllm/model_executor/models/llama.py @@ -278,6 +278,7 @@ def __init__( quant_config: Optional[QuantizationConfig] = None, lora_config: Optional[LoRAConfig] = None, prefix: str = "", + **kwargs ) -> None: super().__init__() self.config = config