I’m trying to use gpt-4-vision-preview for make-up coaching, given an image of the user as an input. That all start with an analysis of some (safe and identification free) facial features.
I’m getting excellent results in ChatGPT Pro Web, but for the exact same prompt and same images, gpt-4-vision-preview performs dramatically worst in terms of analysis.
We are getting a near 90% accuracy on the web version and roughly 50% with the same images dataset and prompt with the API.
I couldn’t find much information on what configuration and model are used in the web version when submitting a prompt with an image attached. Is it the same gpt-4-vision-preview that’s available in the platform API ?
Are the parameters used known / documented, or is there any reliable estimate from the community ? (temperature, etc.)
Thanks in advance for any hint