From e07f21a6d387f0982cfd2aa865c25c372fd373f7 Mon Sep 17 00:00:00 2001 From: Philip Kiely Date: Mon, 19 Aug 2024 14:30:05 -0500 Subject: [PATCH] add max tokens to example --- llama/llama-3_1-8b-instruct/config.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama/llama-3_1-8b-instruct/config.yaml b/llama/llama-3_1-8b-instruct/config.yaml index 755e3840..0425461e 100644 --- a/llama/llama-3_1-8b-instruct/config.yaml +++ b/llama/llama-3_1-8b-instruct/config.yaml @@ -1,7 +1,7 @@ model_name: "Llama 3.1 8B Instruct VLLM" python_version: py311 model_metadata: - example_model_input: {"prompt": "what is the meaning of life"} + example_model_input: {"prompt": "what is the meaning of life", "max_tokens": 1024} repo_id: meta-llama/Meta-Llama-3.1-8B-Instruct tensor_parallel: 1 requirements: