OpenAI has officially rolled out the Advanced Voice Mode (AVM) for all paid subscribers of its popular AI assistant, ChatGPT. This new feature, which has been in development for several months, is now available to users of the Plus and Team plans. The rollout is set to begin in the United States and will gradually expand to other regions, including those using OpenAI Edu and Enterprise plans.
Enhanced Capabilities and Accessibility
The AVM is designed to improve the conversational experience by enhancing the speed, fluency, and pronunciation of conversations in multiple languages. OpenAI has announced that the new feature is now available to all users who have opted for the Plus and Team plans. It will initially be launched in the U.S. market and will be extended to Edu and Enterprise subscribers in the coming week.
New Voices and Customization Options
One of the key improvements in AVM is the addition of five new voice styles: Arbor, Maple, Sol, Spruce, and Vale. These join the existing four voice styles—Breeze, Juniper, Cove, and Ember—giving users nine distinct voice options. OpenAI has also removed the voice style that was previously criticized for resembling Scarlett Johansson’s voice, known as Sky.
Users can now interact with ChatGPT through spoken commands rather than text prompts. This feature is particularly useful for those who prefer a more natural and conversational experience. When entering the voice mode, users will see a pop-up notification confirming their transition to the advanced voice assistant.
Multilingual Support and Custom Instructions
The AVM supports over 50 languages, including a unique feature that allows the assistant to apologize in the user’s preferred language. For example, if a user instructs the assistant to apologize to their grandmother for making her wait, the assistant will first respond in fluent English and then switch to standard Mandarin Chinese upon the user’s request.
This feature not only enhances the user experience but also makes the AI assistant more versatile and adaptable to diverse cultural and linguistic environments. The ability to customize the assistant’s behavior through memory and custom instructions further personalizes the interaction, ensuring that the AI responds according to the user’s preferences.
Security and Functionality
While the new voice mode has many exciting features, it’s important to note that some functionalities from the original demonstration in May have been omitted. For instance, the computer vision feature that allowed the AI to provide voice feedback on users’ dance moves through their smartphone camera has not been included in this update.
Additionally, the new AVM is only applicable to the GPT-4o model, and not the recently released preview model G1. This decision was made to address potential security concerns that were raised during the initial testing phase.
Conclusion
The launch of the Advanced Voice Mode marks a significant step forward in the evolution of ChatGPT. With enhanced language capabilities, new voice styles, and improved customization options, OpenAI is aiming to make its AI assistant more accessible and engaging for a global audience. As the feature rolls out to more regions and users, it will be interesting to see how the AI assistant continues to evolve and adapt to meet the needs of its diverse user base.
Views: 1