diff --git a/vllm/entrypoints/openai/serving_chat.py b/vllm/entrypoints/openai/serving_chat.py index a6fe538dda74f..1dee798858a16 100644 --- a/vllm/entrypoints/openai/serving_chat.py +++ b/vllm/entrypoints/openai/serving_chat.py @@ -135,8 +135,6 @@ async def create_chat_completion( guided_decode_logits_processor = ( await self._guided_decode_logits_processor(request, tokenizer)) - print(request) - print("prompt", prompt) prompt_inputs = self._tokenize_prompt_input( request, tokenizer,