OpenAI has started rolling out its advanced Voice Mode to a select group of ChatGPT Plus users following a brief delay due to safety concerns. The new feature promises more natural, real-time conversations with emotional responsiveness, allowing users to interrupt at any time and sense their emotions.
In an effort to ensure a safe and high-quality voice communication experience, OpenAI conducted extensive testing of GPT-4o’s voice capabilities with over 100 external red teamers across 45 languages. The model has been trained to speak in four preset voices, enhancing the overall user experience.
The company plans to gradually expand the rollout to more users in the coming months, aiming to make Voice Mode accessible to all Plus users by fall. Additionally, OpenAI is set to introduce video and screen-sharing capabilities in the near future, further enhancing the platform’s capabilities.
This development comes on the heels of a delay in the launch of the voice assistant feature for ChatGPT, as OpenAI focused on addressing safety issues. The company worked on improving the model’s content detection capabilities, user experience, and infrastructure scalability to ensure real-time responses for millions of users.
Earlier, OpenAI faced controversy when actress Scarlett Johansson raised legal concerns over the alleged replication of her voice using AI technology. Despite this setback, Microsoft Corp. recently integrated OpenAI models into its Azure cloud platform, highlighting the increasing adoption of advanced AI models in various industries.
The move towards more advanced voice capabilities by OpenAI underscores the growing demand for natural and interactive conversational experiences. As the company continues to refine its technologies and enhance user safety, users can look forward to more engaging and emotionally responsive interactions in the near future.