You can adjust the "temperature" setting to control the randomness of the model's responses. Higher values (e.g., 0.8) make responses more creative, while lower values (e.g., 0.2) make them more focused. You can also use the "max tokens" option to limit the length of the response.
Use Temperature and Max Tokens