diff --git a/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl b/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl index 377b5bce0..eb7b46071 100644 --- a/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl +++ b/benchmarks/inference-server/text-generation-inference/manifest-templates/text-generation-inference.tftpl @@ -51,7 +51,7 @@ spec: - name: text-generation-inference ports: - containerPort: 80 - image: "us-docker.pkg.dev/deeplearning-platform-release/gcr.io/huggingface-text-generation-inference-cu121.2-2.ubuntu2204.py310" + image: "ghcr.io/huggingface/text-generation-inference:1.4.2" # TODO(annapendleton,alvarobartt): update to Hugging Face DLC once verified args: ["--model-id", "${model_id}", "--num-shard", "${gpu_count}", "--max-concurrent-requests", "${max_concurrent_requests}"] env: %{ for hugging_face_token_secret in hugging_face_token_secret_list ~}