Hi there
Just found so many interesting ideas on OpenAI community.![]()
Like many others, I’ve been thinking about the better way integrating AI with our daily work. Thinking about how we can best utilizing the power of LLM like GPT-4o.
A couple month later, may be a year, after being use ChatGPT and GPT-4 API, and some other third-party Copilot AI - like client in my daily work. I started to realize that how natural it is when you can talk to your working space(the screen) directly with AI anytime, with just a screenshot taking. I mean visually talking to, not copy-pasting text.
Along with the recent release of GPT-4o, which beat GPT-4 with more powerful multimodal capability and lower price. I think it is time to create such a copilot app. That’s how MacCopilot come from.
How to use
- Take screenshot by keyboard shortcut;
- Input the question.
That’s it, thanks to GPT-4o, you can talk to your whole screen.
Show case
It is a paid app with no free tier for now. You also need a OpenAI API key that can use GPT-4o model to use it.
You can visit maccopilot.co for more information, or buy it on Gumroad - MacCopilot.
Other features
Beside gpt-4o, gpt-4-turbo, gpt-4-vision, MacCopilot also support other platform like Claude Opus, Gemini AI. Azure OpenAI.
Plans
- A ready-to-use subscription with builtin API key.
- Integrate open-source multimodal models.
Hope it can help others think about more forms of Copilot AI.
Please feel free to share thought below.
