From 92055dedfd4b26e0b377a7e6238b0c68fb2ec2b4 Mon Sep 17 00:00:00 2001 From: Luiz Oliveira Date: Wed, 6 Mar 2024 09:13:06 -0500 Subject: [PATCH] Bump TGI version to 1.4.2 to support gemma models (#288) --- .../manifest-templates/text-generation-inference.tftpl | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl b/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl index 7077af515..01e613828 100644 --- a/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl +++ b/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl @@ -50,7 +50,7 @@ spec: - name: text-generation-inference ports: - containerPort: 80 - image: "ghcr.io/huggingface/text-generation-inference:1.1.1" + image: "ghcr.io/huggingface/text-generation-inference:1.4.2" args: ["--model-id", "${model_id}", "--num-shard", "${gpu_count}"] # , "{token}" tensor parallelism, should correspond to number of gpus below %{ for hugging_face_token_secret in hugging_face_token_secret_list ~} env: