It’s great that OpenAI keeps pushing forward with new models, but before focusing on bigger and more powerful AI, the platform itself needs serious improvements. There are core functionality issues that make working with AI frustrating and inefficient.
Here are the biggest problems that need attention right now:
1. System Prompt Generator is Useless
- The chatbot that “helps” write system prompts spits out unstructured, blog-style text instead of a clear, formatted prompt that actually works.
- It doesn’t follow best practices for AI prompting, making it more work to fix than to just write it manually.
- No clear sections, variable placeholders, or structured formatting—just a wall of text.
2. PDFs and Document Formatting is Broken
- AI never generates PDFs correctly—formatting gets destroyed, spacing is inconsistent, and text alignment is all over the place.
- It randomly removes or alters details, making it unreliable for professional use.
- It doesn’t recognize proper document standards, so anything AI-generated looks amateurish.
3. Code Generation Still Repeats the Same Mistakes
- AI forgets context when generating code, often missing dependencies or previous variables.
- No consistency in formatting—sometimes it writes clean, structured code, other times it’s messy.
- It doesn’t debug properly—if a user points out an issue, AI often suggests the same broken fix over and over.
4. System Prompts Get Compressed Without Warning, Causing Debugging Issues
- Users were never informed that long system prompts get compressed into a directive, which means important details get lost.
- There’s no way to see how the system prompt is actually being interpreted, leading to unpredictable GPT behavior.
- If users try to debug by asking how the system prompt is being processed, GPT often refuses to answer, assuming it’s a memory extraction attempt.
- This makes fine-tuning and debugging nearly impossible, forcing users to guess how the AI is applying their instructions.
5. AI Doesn’t Handle Multi-Step Tasks Well
- If a request involves multiple steps (e.g., “Generate an Excel file, then analyze the data, then write a report”), it often fails partway through.
- It doesn’t properly chain actions together, forcing users to re-explain things over and over.
- Memory resets mid-task, making it unreliable for workflows that require context retention.
6. The Canvas Tool is a Mess
- AI often writes over existing content in Canvas, leading to lost work.
- It will say it edited something, but the changes never actually apply.
- The tool hallucinates edits, claiming it adjusted something when it didn’t.
- Overall, Canvas feels unreliable and is more of a frustration than a useful tool.
7. AI Tools (like Browser, Python, etc.) are Unreliable
- Web searches frequently return useless or outdated info, sometimes even hallucinating results.
- The Python execution tool times out too easily, making it unreliable for longer computations.
- File handling is inconsistent—sometimes AI can extract data from files, other times it just fails with no explanation.
OpenAI should fix these core issues first before focusing on new models. These are real pain points that make AI harder to use than it should be. A more powerful model won’t solve these problems if the platform itself is broken.
Has anyone else been frustrated by these issues? What else needs fixing?