Temperature, top_p and top_k for chatbot responses

softmax temperature can be though of as the amount of noise injected into the decision-making process

top-p can be considered a weighting that pushes more towards selecting top results

The current models don’t need the temperature increased to be “creative”, they already produce poorer tokens than before. Increasing will only help to break deterministic output for you on repeated runs.