-
-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Frontend] Gracefully handle missing chat template and fix CI failure #7238
[Frontend] Gracefully handle missing chat template and fix CI failure #7238
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge). To run full CI, you can do one of these:
🚀 |
…vllm-project#7238) Co-authored-by: Roger Wang <[email protected]>
…vllm-project#7238) Co-authored-by: Roger Wang <[email protected]>
…vllm-project#7238) Co-authored-by: Roger Wang <[email protected]>
…vllm-project#7238) Co-authored-by: Roger Wang <[email protected]> Signed-off-by: Alvant <[email protected]>
…vllm-project#7238) Co-authored-by: Roger Wang <[email protected]>
This PR fixes the failing async engine test which contains invalid examples as of transformers v4.44 (see huggingface/transformers#31733).
This PR also improves the error handling when the tokenizer does not contain a chat template before calling
tokenizer.apply_chat_template
, since the error raised inside HuggingFace isn't as relevant to the user.