Does the OpenAI Real-Time API natively support text-to-speech (TTS) and speech-to-text (STT) functionalities, or do we need to configure tools like Whisper and TTS voice models manually using WebSockets?
Does the OpenAI Real-Time API natively support text-to-speech (TTS) and speech-to-text (STT) functionalities, or do we need to configure tools like Whisper and TTS voice models manually using WebSockets?
Yes, it handle it by itself
Thanks @kevin6 It is super helpful answer)