diff --git a/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java b/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java index 65c87398a..d5dc27a07 100644 --- a/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java +++ b/model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java @@ -30,7 +30,7 @@ public interface ChatModelConfig { * An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens * with topP probability mass. * 0.1 means only the tokens comprising the top 10% probability mass are considered. - * It is recommended to alter this or topP, but not both. + * It is recommended to alter this or temperature, but not both. */ @WithDefault("1.0") Double topP();