This is a great post by Lysanne van Beek about the misconceptions and surprises she encountered during our Intro to AI c
This is a great post by Lysanne van Beek about the misconceptions and surprises she encountered during our Intro to AI course!
One of the surprises is that ChatGPT never gives the same answer twice, and Lysanne goes a bit into the topic, but it’s actually much more fascinating!
LLMs have the concept of temperature, which basically says how creative the model should be with its responses. High temperature translates to higher creativity.
However, even if we set the temperature to 0—instructing the model to always pick the most probable word next—the model will give different answers to the same prompt.
Why?
The reason is… speed! While the details are complicated, the general idea is as follows: to provide an answer quickly, the computation is distributed across different machines. These machines can round floating-point numbers slightly differently because the values stored in memory are slightly different. Small differences can create significantly different probabilities, meaning… You never get the same answer!
I found this part really impactful, as even though you set the temperature to 0, models cannot offer reproducibility.
This makes having a solid evaluation framework even more important to ensure the model continues to behave in the desired way.
https://xebia.com/blog/7-things-that-always-surprise-people-in-our-intro-to-ai-course/
