Interface ChatModelConfig
public interface ChatModelConfig
-
Method Summary
Modifier and TypeMethodDescriptionformat()the format to return a response in.Whether chat model requests should be loggedWhether chat model responses should be loggedMaximum number of tokens to predict when generating textseed()With a static number the result is always the same.stop()Sets the stop sequences to use.The temperature of the model.topK()Reduces the probability of generating nonsense.topP()Works together with top-k.
-
Method Details
-
temperature
The temperature of the model. Increasing the temperature will make the model answer with more variability. A lower temperature will make the model answer more conservatively. -
numPredict
OptionalInt numPredict()Maximum number of tokens to predict when generating text -
stop
Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return -
topP
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text -
topK
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative -
seed
With a static number the result is always the same. With a random number the result varies Example:Random random = new Random(); int x = random.nextInt(Integer.MAX_VALUE); -
format
the format to return a response in. Currently, the only accepted value isjson -
logRequests
@ConfigDocDefault("false") @WithDefault("${quarkus.langchain4j.ollama.log-requests}") Optional<Boolean> logRequests()Whether chat model requests should be logged -
logResponses
@ConfigDocDefault("false") @WithDefault("${quarkus.langchain4j.ollama.log-responses}") Optional<Boolean> logResponses()Whether chat model responses should be logged
-