タイトル(Title)
“Japanese TTS sometimes misreads kanji – is your Japanese NLP team not talking to the TTS team?”
内容(Body)
Hi, I’ve noticed a consistent issue in the Japanese voice output (TTS):
- Even when the kanji is correct, the pronunciation is sometimes wrong.
- The same kanji might be read correctly in one sentence but misread in another.
- It seems like the reading (yomigana) is not contextually tied to the generated text.
I’m wondering:
- Is the TTS module running separately from the LLM?
- Are the pronunciation rules handled independently from the Japanese text generation team?
- Or… are the two teams just not getting along?
On a serious note: for Japanese users, accurate pronunciation is critical for usability — especially when it comes to uncommon kanji or names. Even if humans can mentally correct it, it breaks immersion.
Thanks for the great product overall! Just wanted to nudge the TTS harmony a bit
「This is in reference to ChatGPT’s Japanese voice mode (advanced and default)」