.. | ||
src/main | ||
pom.xml | ||
README.md |
Demo 02 - LLM configuration
In this step, we will play with various configurations of the language model (LLM)
Temperature
quarkus.langchain4j.openai.chat-model.temperature
controls the randomness of the model’s responses.
Lowering the temperature will make the model more conservative, while increasing it will make it more creative.
Max tokens
quarkus.langchain4j.openai.chat-model.max-tokens
limits the length of the response.
Frequency penalty
quarkus.langchain4j.openai.chat-model.frequency-penalty
defines how much the model should avoid repeating itself.