Thanks for replying to this thread, I just can’t seem to work out the Vision capabilities in the Playground so I was wondering maybe I don’t have it at all, since the documentation says here:
GPT-4 with Vision, sometimes referred to as GPT-4V or gpt-4-vision-preview in the API, allows the model to take in images and answer questions about them.
GPT-4 with vision is currently available to all developers who have access to GPT-4 via the gpt-4-vision-preview model and the Chat Completions API which has been updated to support image inputs.
Since my model dropdown doesn’t explicitly say “GPT-4V” or “gpt-4-vision-preview”, do I even have that option?
Yes, you absolutely have access to vision if you have gpt-4. As of writing this post, the playground only supports text inputs for chat completion and completion API.
Hence, the best way to try the vision model is to simply use the list models endpoint to check the models available to you and use the boilerplate code in the documentation I shared to access the vision model.