From 31226a0b2f4332153bfdc1cbad5fd18778125c56 Mon Sep 17 00:00:00 2001 From: Pierre Beitz Date: Thu, 19 Sep 2024 15:43:19 +0200 Subject: [PATCH] [doc] Fix a typo in the ChatModelConfig --- .../langchain4j/openai/runtime/config/ChatModelConfig.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java b/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java index 65c87398a..d5dc27a07 100644 --- a/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java +++ b/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java @@ -30,7 +30,7 @@ public interface ChatModelConfig { * An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens * with topP probability mass. * 0.1 means only the tokens comprising the top 10% probability mass are considered. - * It is recommended to alter this or topP, but not both. + * It is recommended to alter this or temperature, but not both. */ @WithDefault("1.0") Double topP();