Skip to content

Commit

Permalink
[V1] Fix Detokenizer loading in AsyncLLM (vllm-project#10997)
Browse files Browse the repository at this point in the history
Signed-off-by: Roger Wang <[email protected]>
  • Loading branch information
ywang96 authored and weilong.yu committed Dec 13, 2024
1 parent 063db3c commit 2f08836
Showing 1 changed file with 6 additions and 1 deletion.
7 changes: 6 additions & 1 deletion vllm/v1/engine/async_llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,12 @@ def __init__(
input_registry)

# Detokenizer (converts EngineCoreOutputs --> RequestOutput).
self.detokenizer = Detokenizer(vllm_config.model_config.tokenizer)
self.detokenizer = Detokenizer(
tokenizer_name=vllm_config.model_config.tokenizer,
tokenizer_mode=vllm_config.model_config.tokenizer_mode,
trust_remote_code=vllm_config.model_config.trust_remote_code,
revision=vllm_config.model_config.tokenizer_revision,
)

# EngineCore (starts the engine in background process).
self.engine_core = EngineCoreClient.make_client(
Expand Down

0 comments on commit 2f08836

Please sign in to comment.