The May demo of GPT-4o showcased its impressive speech recognition and translation skills, as well as its ability to understand and respond to emotions and nonverbal cues. It demonstrated natural voice conversations, real-time audio translation, and enhanced multilingual and audiovisual processing capabilities.
The new ChatGPT features, including voice and emotion-reading capabilities, are expected to be available to all paying users in the fall, according to OpenAI. However, the company emphasizes that exact timelines will depend on meeting their high safety and reliability standards.
OpenAI planned to add voice and emotion-reading features to its ChatGPT chatbot, enabling more natural conversations with AI. These features would allow the chatbot to respond to a person's tone of voice and facial expressions, and have more complex conversations with expressive tones.