Announcing GPT-4o in the API!

Hi Thank you for the work you do Team OpenAI, this tool will change the world. I hope to test the audio and video module soon. I would love to see its implementation in different types of developments, especially in mods with video games, experiments xD.

1 Like

Pretty cool, I’m really hoping the audio features in the API comes soon, it will be perfect for my enterprise!

The multimodal capability is insanely great! :blush:

Have you tried it yet or are you just basing that off the demo?

1 Like

I do tried it for a while. Indeed I developed an app for macOS utilizing the multimodal capability of gpt-4o, just realizing the potential of it.

The text to speech and speech to text for Omni model aren’t available yet. Same thing for Omni model generating images. Did you just use the models that were already available or do you have beta access to all of Omni models capabilities?

No,none of these two available yet. In my case, only text to text and image to text is used for now. The app were used to interact with the screen content.

Brandon W via OpenAI Developer Forum <notifications@openai1.discoursemail.com>于2024年6月13日 周四22:11写道:

yet

So you are doing nothing more than everyone else.

Got excited there for a second. Shame!

Well, it is not my point confusing you. Developing the app was to address a specific problem effectively, and I believe it does well.

Regarding sharing links, it seems the community prevent posting them directly. You can search for “MacCopilot” and give it a shot if you are interested.

Congrats! That push me into the new gen of IA!! :upside_down_face: