all AI news
How does temperature impact next token prediction in LLMs?
Towards Data Science - Medium towardsdatascience.com
TLDR
1. At a temperature of 1, the probability values are the same as those derived from the standard softmax function.
2. Raising the temperature inflates the probabilities of the less likely tokens, thereby broadening the range of potential candidates (or diversity) for the model’s next token prediction.
3. Lowering the temperature, on the other hand, makes the probability of the most likely token approach 1.0, boosting the model’s confidence. Decreasing the temperature effectively eliminates the uncertainty within the model. …
deep learning diversity function impact large language models llms next prediction probability python softmax softmax-function standard token tokens values