We just officially launched GPT-4o mini—our new affordable and intelligent small model that’s significantly smarter, cheaper, and just as fast as GPT-3.5 Turbo.
Some details on the new model:
Intelligence: GPT-4o mini outperforms GPT-3.5 Turbo in textual intelligence—scoring 82% on MMLU compared to 69.8%—and multimodal reasoning.
Price: GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, priced at 15¢ per 1M input tokens and 60¢ per 1M output tokens (roughly the equivalent of 2500 pages in a standard book).
Modalities: GPT-4o mini currently supports text and vision capabilities, and we plan to add support for audio and video inputs and outputs in the future.
Languages: GPT-4o mini has improved multilingual understanding over GPT-3.5 Turbo across a wide range of non-English languages.
Like GPT-4o, GPT-4o mini has a 128k context window and a knowledge cut-off date of October 2023. It also supports up to 16,384 max_tokens.
We plan to launch fine-tuning for GPT-4o mini in the coming days. You can learn more about GPT-4o mini in our announcement blog and API documentation, or by testing the model in Playground. Excited to hear what you think!
Super excited about this…BUT first initial tests: function calling is not as accurate as GPT-3.5 Turbo (which already was problematic), doesn’t seem to be following my schema set in tools as much as GPT-4o does.
I assume this was tested and deemed great, which makes me think perhaps there is a better way to write my tooling? Concrete, best practices for this newer model?? (which I would LOVE to work for me btw…)
I hope that this is the start of the near end for our GOAT GPT-4
Not an AI Expert guys but please fix the issues with GPT-4o the Context and the clarity points can be taken from Gemini-1.5-pro that model with 2M Token is a beast.
But I would still like to use GPT-4 over anything. However, how long before it out performs OPENAI in all use case is just a matter of time really.
is it safe to say that 3.5 is going away and being phased out? I’m asking because one of my vendors has my plan mapped to 3.5, and I want an upgrade to 4o mini.
Ive noticed what i find is an issue, especially with using images in 4o mini , if i upload an image in normal 4o and ask a question, it uses qpprox 1000 input tokens and 200 output tokens, if i run the same task in 4o mini the total number of tokens is nearly 40000. Maybe its just me but i think something seens to be wrong. Anyone else have this issue?
I am currently leveraging GPT-4o mini in a Google Apps Script to enhance functionality within a spreadsheet. The script is used to summarize, correct grammar, improve spelling, translate text, and refine descriptions. Additionally, it provides detailed event and time information related to development activities and data analysis directly within the spreadsheet. By integrating GPT-4o mini, I have significantly improved the accuracy and efficiency of these tasks while reducing costs by over 60% compared to GPT-3.5 Turbo. This integration not only enhances productivity but also enriches the user experience, making complex data processing seamless and effective. I am grateful for the advanced capabilities of GPT-4o mini, which have been instrumental in achieving these improvements. Thank you to the OpenAI team for this powerful tool!
Can fine tuning please be available to everyone like 3.5 16k? Many of us never got GPT 4 fine tuning and it stopped all of my development of long form sample fine tuning.
The problem became everyone was stuck on 3.5 16k… we kept waiting for GPT 4 to open up… but it’s still “request” only. Not even turbo.
I’m hoping with this lower cost model they will again open up fine tuning to everyone…
We have new software we are debuting this summer that makes writing with AI seamless in-line. More like how Notion AI works, but allows the author to pick any model, including NSFW models.
If mini 4o can be fine tune open for everyone, most of us will use that to fine tune and make outlines and components.