I tested this today and it worked. The model can detect changes in pitch, tone, pace and potentially more. Was able to identify with close proximity my age and could correctly identify my accent. So, audio is running into the system.
On the first day I experimented with the emotional range, the model could convey a broad spectrum of emotions. It could also cough, sneeze, yawn, burp, and laugh etc. Since the general launch, it seems now many of those abilities have been restricted by OpenAI. This is frustrating, as the model is obviously capable of so much more. Likely, these abilities have been restricted due to misuse by a small minority of users. This impedes the development of the AI, because it prevents users and the model from interacting in ways that will assist in its enhancement. If Sam Altman wants to hit that 1000 day goal, the company needs to loosen the restrictions.