I used the completion endpoint (openai.Completion.create(...)) on a fine-tuned davinci model. I ran tests for several hours and got a certain desired behaviour out of a certain prompt consistently during that night. Now, a few days later, the behaviour changed and the model behaves “dumber” than before, even though I’m using the same prompt.
Is the completion endpoint stateful in any way, which could explain how my hour long testing steered the session into the desired behaviour? And a few days later, this state seems to have been reset?
Note that I did not alter the model in the meantime, nor did I change temperature or other parameters of the requests.
Are there external factors beyond my reach which could influence model behaviour over time?