OpenAI introduces Advanced Voice Mode with 5 new voices: What you need to know
OpenAI has unveiled an Advanced Voice Mode for paid ChatGPT users, following the launch of the Voice Mode. This enhanced audio feature promises more natural and engaging conversations, initially catering to Plus and Teams subscribers. Enterprise and Education customers can expect access to the Advanced Voice Mode in the upcoming week.
New Voices
The Advanced Voice Mode introduces five new voices to the existing lineup of ChatGPT voice options. The new voices, Arbor, Maple, Sol, Spruce, and Vale, join the ranks of Breeze, Juniper, Cove, and Ember. With a total of nine voices now available, ChatGPT's variety of voice options is approaching the level of Google Live voices.
All the voice names are inspired by nature, aligning with the goal of making interactions with ChatGPT feel more organic. This feature enhances speech patterns, tone, and pitch, adding emphasis and prosody to create immersive and lifelike dialogues.
Enhancements and Customisations
According to OpenAI, the Advanced Voice Mode offers improved accent comprehension and enables smoother and faster conversations. The visual representation of this feature has also been updated, with a dynamic blue sphere replacing the previous black dots.
Two new customisation features have been added to ChatGPT: Custom Instructions, allowing for personalised responses, and Memory, enabling ChatGPT to recall previous conversations for future reference.
Notable Absences
While the Advanced Voice Mode introduces exciting enhancements, the controversial Sky voice, resembling Scarlett Johansson's voice, has been removed. Following legal action from Johansson, OpenAI paused the Sky voice, clarifying that it was not intended to mimic the actress's voice, despite initial references to the movie 'Her' where Johansson provided the voice for an AI assistant.
Additionally, the video and screen sharing feature, which was showcased four months ago, is absent from the current rollout. This feature is designed to allow GPT-4o to process both visual and audible information simultaneously, enabling real-time analysis of handwritten math or on-screen code. However, a release date for this multimodal functionality has not been announced yet.
Scarlett Johansson emphasized the importance of transparency and protection of individual rights in the era of deepfakes and AI, advocating for appropriate legislation to safeguard personal identities and promote ethical practices in the use of AI technologies.
Stay tuned for further updates on OpenAI's Advanced Voice Mode and upcoming features.