Skip to content

Commit

Permalink
[Misc] improve logits processors logging message (vllm-project#7435)
Browse files Browse the repository at this point in the history
  • Loading branch information
aw632 authored and kylesayrs committed Aug 17, 2024
1 parent bd8702c commit 33e8a91
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 2 deletions.
1 change: 0 additions & 1 deletion tests/entrypoints/openai/test_serving_chat.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,6 @@ def test_serving_chat_should_set_correct_max_tokens():
with suppress(Exception):
asyncio.run(serving_chat.create_chat_completion(req))

# AsyncLLMEngine.generate(inputs, sampling_params, ...)
assert mock_engine.generate.call_args.args[1].max_tokens == 93

req.max_tokens = 10
Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/logits_processors.py
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ def get_logits_processors(
# Check if token_id is within the vocab size
for token_id, bias in clamped_logit_bias.items():
if token_id < 0 or token_id >= tokenizer.vocab_size:
raise ValueError("token_id in logit_bias contains "
raise ValueError(f"token_id {token_id} in logit_bias contains "
"out-of-vocab token id")

logits_processors.append(
Expand Down

0 comments on commit 33e8a91

Please sign in to comment.