Same instructions, different results: GPT vs OpenAI Assistant (message splitting issue too)

Hi everyone,

I’m noticing a big difference in behavior between a custom GPT and an OpenAI Assistant — even though I’m using exactly the same instructions and files in both.

The GPT (even using gpt-4o-mini) follows the instructions well and produces responses that align with what I expect. However, when I test the same setup using the Assistant (via Playground), the results are very different — even after trying different models and low temperatures (0 to 0.2).

One major issue is message splitting. I’ve been very explicit in my system instructions, repeatedly stating things like:

“Never delay or split your response. Always send full replies (greeting, policy, and content) as one single message.”

and:

“Always respond with a single, complete message — never divide into multiple parts. Do NOT break the response into separate replies under any circumstances — including the first interaction.”

Despite this, the Assistant continues to split replies, especially on the first interaction — which breaks the expected flow, particularly for WhatsApp-based use.

Has anyone else run into this? Is there something I’m missing about how the Assistant handles instructions compared to a custom GPT?

Any help would be much appreciated!