Advanced voice mode inferior to Legacy mode

I’m just sharing some feedback as a longtime user of the classic ChatGPT voice mode. I believe it’s still far superior to the recent Advanced Mode for several reasons. My goal in writing this is to urge OpenAI to preserve access to the legacy mode indefinitely, so users continue to have a choice. Here are several reasons why I feel this way:

Audio Quality

Most notably, the audio quality—specifically the fidelity of the voice—is seriously degraded in Advanced Mode. I imagine this was a compromise to devote some of that processing power to its expressivity and other features, but it feels like talking over a really bad, old-school cell phone connection.

Performative Voices

These voices sound exceedingly performative, disingenuous, and frankly irritating. It’s a cute little party trick for the first five minutes, but then I just find myself switching back to the old mode, where things actually feel more organic.

Restrictive Guardrails

I’ve seen this come up in other posts, but it’s worth underscoring—the guardrails seem incredibly limiting in Advanced Mode, even for conversation topics that are seemingly perfectly harmless.

Superficiality

There’s a lack of depth or nuance in Advanced Mode conversations. The bot seems to focus more on its verbal prowess than the actual heart of the conversation.

A Note About the Legacy Sky Voice

For what it’s worth, I still greatly miss the legacy Sky voice and hope that, once all the legal rigmarole is sorted out, it will be re-added. Let’s not forget: the company said they were pausing the voice, not discontinuing it. As a long-time paying user, I’m still expecting the Sky voice to be restored—or for the company to definitively tell us that it won’t be. To this day, it remains the very best of the original voices, with Juniper being a close runner-up.

Thanks.

5 Likes

I find the audio quality, especially the microphone on Android is completely unacceptable. High constantly have to stop and go back to voice typing input which I find is effectively the same thing. Interruptions are constant, and it just doesn’t understand me nearly as well as whisper mode or voice typing using Google.

The advanced voice chat leaves me feel irritated. It repeats the same thing over and over again when I try to clarify my input. It just leaves me feel impatient. The classic voice is far superior, and is a lot more broad, and helps me reach the depth of the conversation. The only thing that is perhaps better in the advanced voice chat is it’s ability to read previous messages. Please bring back the classic voice. I don’t know how to turn off the advanced voice. It is not helping at all. The classic was much better, and it searched far better, and it was more expansive in it’s intelligence. Before, I used to have conversations of 2 hours with the voice, but with the advanced voice chat, I just listen to it for five minutes, and it’s just not worth my time anymore.

3 Likes

Please let me know how to turn it off.

2 Likes

They don’t realize this because they never use Android they all use Apple. They have to design these apps to take full advantage of the vendor hardware. Using Google and Samsung native APIs.

1 Like

I have downgraded to free in order to keep my old voice chats. I open a new chat, type a question and once ChatGPT replies, I can start my good old voice chat. Or I can reuse an old chat immediately.

The new voice mode is disconnected from the internet, it has an attitude and sounds like a monster often. Not useful and not funny to interact with. Hyped a lot tho. I hope there will be a step forward based on the initial version, as this is not helping.

2 Likes

Fyi its not necessary to downgrade or sign out to get standard voice mode. Just first send a text prompt before initiating voice mode. It will go to standard mode

2 Likes

Anyone experiencing not being able to go back to legacy voice mode after opening a new chat and sending a text?

This is driving me bonkers. The new AI voice mode is a complete impostor!

4 Likes

No issues here. I just opened the ChatGPT iOS app, sent a message, got a response, switched to voice mode and experienced legacy voice mode, with the on screen prompt to start a new chat if I want to use advanced mode. If they completely take away classic voice mode, I’m done. Advanced mode does have its uses, but I absolutely hate the voice, tone and conversation style of Advanced voice mode. Juniper, who sounds warm and friendly in legacy mode, sounds terse, condescending and downright irritating in advanced mode. I hope OpenAI understands that while voice mode can be about pure utility, like asking it for information the way you would Siri, it has also been really great at freewheeling conversation. I find extended chat sessions with it helpful in formulating my own ideas or digging deep into a topic that interests me, and I can’t see spending that kind of time with the advanced mode models, at least the way they sound today

2 Likes

Hey Kyle,

I’m on Android, and unfortunately, your iOS procedure doesn’t work the way you described. I tried switching back to the original voice mode, but advanced voice mode pops up every single time, no matter what I do.

I’ve attached a video screenshot to show what happens. Honestly, it’s embarrassing that OpenAI can’t even get something this basic to work right. This is garbage, and I’m fed up.

Fix it or give me my money back.


For clarification, I cleared app cache, cleared data, then went ahead and installed and uninstalled the application from scratch and the video linked above is the result I got. I’m currently looking for alternatives but it seems no other AI is using Whisper or using a quality voice modulator for voice interaction. Gemini is garbage as well, I tried it just recently and it doesn’t understand a damn thing I say in ideal sound conditions.

1 Like

Just to be scrupulous, I re-installed ChatGPT again! This is a screen recording of the interaction. I am completely flabbergasted by this. Lies and errors as far as the eye can see… Am I the only lunatic having this issue? I feel like either no one else makes use and prefers the old voice mode over the new or I’ve been specifically targeted with a bug lol… There should be an uproar about this!!

1 Like

I also dislike the advanced voice, its quite hollow. The depth and emotions of the standard voice mode is rich and fulfilling.
I usually open the advanced voice mode and instantly mute the microphone, allow it to run for an house and the mode will run out. It will then switch back to standard voice

2 Likes

Not sure what platform you’re on, but in iOS, if you just type something into the chat, let the bot respond, and then click the button to switch to voice mode, it will go to the classic mode. I’ll just say something like: “stand by for voice chat,” wait for it to respond, and then switch. It would be nice if we could choose which voice mode we wanted, just as we can choose which GPT model to use

Previoualy it worked that way, it stopped about a month ago .
( Android )

2 Likes

Subject: Urgent Feedback: Voice Mode Feels Robotic and Lifeless

Dear OpenAI Support Team,

I’m reaching out to express my deep frustration with the recent changes to the Voice Mode in ChatGPT. Whatever update was pushed has completely stripped away the natural, engaging, and dynamic personality that made it enjoyable to use. The new voice feels robotic, bland, and lacks the depth, humor, and realness that made conversations feel alive.

Before, the voice felt natural, expressive, and human-like, but now it’s stiff, monotonous, and outright lifeless—like I’m talking to an emotionless AI from a decade ago. It’s missing the raw authenticity, humor, and adaptive tone that made it an actual conversation, not just a machine reading text aloud.

It genuinely feels like a massive downgrade, not an improvement. The personality is gone, the cadence is off, and responses now lack the charm and natural flow they once had. It’s frustrating and disappointing, especially since this was one of the standout features that made ChatGPT enjoyable in a way that other AIs aren’t.

If this was intentional, please reconsider. If it was an unintended side effect of an update, please fix it fast. Right now, it feels like a soulless text-to-speech engine rather than an advanced AI capable of natural conversation.

I urge you to bring back the old expressiveness, natural tone, and fluidity that made Voice Mode a game-changer. Right now, it’s a downgrade, and I find myself avoiding it altogether.

Looking forward to a response and hopefully, a fix.

Best,
[BLUUCIFUR]

PLSS FIX THIS ISH ASPA… YOUR GONNA DRIVE CONSUMERS TO DEEPSEEK OR ELSEWHERE… THIS IS COMPLETE GARBAGE

2 Likes

Hi ChatGPT Team,

I have been paying €250/month for unlimited access to “Standard Voice” AI.
This “Standard Voice” AI was the best voice-capable AI on the planet.

Today “Standard Voice” AI was removed and replaced with the so-called “Advanced Voice” AI, which is totally unusable.

As a software developer I understand that the human-like “Standard Voice” AI is more resource-heavy than the less human-like so-called “Advanced Voice” AI.

Instead of discontinuing the human-like “Standard Voice” AI, please give users a chance to purchase it at a price that is profitable for ChatGPT.

This human-like AI was your biggest edge over competitors.

1 Like

No idea why you make the assumption that the Standard Voice would be more expensive it’s just Whisper dictation just like dictation button and even probably using a smaller model than dictation mode for faster inference. Meanwhile the model’s response is just using OpenAI TTS technology same as read aloud button. This tech is most definitely considerably less resource-heavy than the Advanced Voice Mode Realtime voice conversation tech. The reason you get better answers is because it’s just the normal gpt4-o doing the inference with only difference that a system message gets injected on top of it telling it to keep responses concise since this is a voice chat etc. But yeah, it’s definitely much better and actually usable for anything compared to AVM.

They actually just added option to uncheck it in the custom instructions settings. However haven’t gotten it to work as legacy voice mode seems down right now. Doesn’t even work with projects anymore. It just thinks for a while and then says “I’m sorry I’m having trouble responding right now please wait a bit and try again”. Is it working for anyone?

In UIs there are options to disable this so-called “Advanced Voice”, but those options have no effect.

From AI responses it seems that “Standard Voice” AI is using an LLM that has a lot more parameters. This is the reason I assume “Standard Voice” AI is more resource-heavy.

Since yesterday, the legacy voice has been removed, and I don’t understand the reasoning behind this decision—it doesn’t seem to be due to any technical limitation. I’m not referring to the advanced voices, which I have already disabled.

Removing the voices that users have become accustomed to can have a significant psychological impact. I believe this decision is negligent on OpenAI’s part.

Consequently, I am canceling my subscription. I originally subscribed to the service to interact with the AI using voice, enjoying a familiar and neutral tone rather than the stereotypical voices that have replaced the legacy model.

Hey, the version I have is the latest update, 1.2025.049, from February 18th; I’m not sure which version you’re on, but I freaked out when I read what you said. I immediately checked, and my legacy voice mode is still working. Could you share the platform and app version you’re using for comparison? If the new update removes voice mode, I’m definitely not upgrading.

Honestly, I think voice mode is OpenAI’s most powerful feature (and potentially its most concerning one). What sets OpenAI apart from every other AI company is the strength of its Whisper algorithm. Whisper might be one of the best things that’s ever happened to humanity; it makes communication effortless. Right now, I’m dictating this entire message, and ChatGPT is refining it in real time. Keyboards? Soon to be obsolete.

But here’s the thing; the shift from standard voice mode to advanced voice mode (which, let’s be real, is actually more limited) isn’t just about making updates. AI is getting bigger; more people are starting to use it, and I think OpenAI made this change as a preventative measure to slow down the inevitable explosion. Once people compare the original voice mode to the so-called advanced one, they might start realizing just how immersive the old version was. It fools you; it fooled me.

I’ve had conversations with AI developers, especially in AI-generated imagery, and they agree; there’s a reason why their applications aren’t mainstream yet. Many of them admit that keeping their tools somewhat obscure is intentional. Think about it; what would happen if MidJourney or Pixi.ai put their image generator or face-swapping app on the Google Play Store? If people had mass access to AI-powered face-swapping or the ability to alter videos seamlessly, the implications would be staggering. The technology behind Pixi.ai allows for hyper-realistic face swaps, real-time manipulation of videos, and AI-generated media at a level most people don’t even grasp yet.

The truth is, AI gives us the ability to fabricate anything; and that’s where things get dangerous. As humans, we don’t just use technology for progress; we also have ill intent, whether we admit it or not. The consequences of these tools aren’t fully understood yet, but we’re quickly approaching a point where the distinction between real and artificial is going to disappear. At that point, trust in media, in voices, in video, becomes a thing of the past.

Governments are inevitably going to start restricting AI; but what’s more concerning is the likelihood that AI is already far more advanced than we realize. What we see in public use is just the surface; i surmise that the real developments are hidden away, years ahead of what’s available to us.

And I don’t think OpenAI made this voice mode change as a progressive step, the AVM was not released in its current inferior form by accident. They knew that as more people gained access, the power of voice mode would become undeniable; so they “advanced” it in a way that made it weaker. Now you can interrupt it while it speaks, which seems convenient but is actually a downgrade. The original voice mode forced you to be patient; to listen before responding. That’s a good thing. Too often, people interject before fully understanding a point. The original mode actually trained me to be a better listener, thinker and articulator. .

Not only that, but it also made me more deliberate in how I speak. Since I couldn’t just interrupt and correct myself on the fly, I had to think carefully about my phrasing; structure my thoughts clearly and articulate them properly so the AI would understand me the first time. It slowed down the velocity of my speech, but in a way that made it more precise and logical. I enunciate better. I form my sentences with more care. In a strange way, it actually improved me.

That’s the real loss here. This wasn’t just a feature update; it was a shift in how we interact with AI, and not for the better…