OpenAI has launched the new Voice Mode for a small group of ChatGPT Plus users. Learn about the features available and how to access Voice Mode in alpha.
- The much anticipated Advanced Voice Mode (AVM) has been released by OpenAI. It is accessible to a limited number of ChatGPT Plus members, though.
- It is anticipated that the new version will outperform the previous one. However, alpha will only offer a restricted set of functionality.
The company that created ChatGPT, OpenAI, showcased the new Voice Mode—which makes use of GPT-4o’s audio and video capabilities—during its Spring Launch event in May. The much anticipated enhanced Voice Mode has been released by the artificial intelligence (AI) research company. On the other hand, only a select few can access it.
The business revealed in an X post that it would be providing ChatGPT Plus subscribers with a more intelligent voice assistant that could react to emotions and be interrupted by introducing Voice Mode in alpha.
What Is Voice Mode?
Voice Mode is an intelligent virtual assistant that enables users to converse back and forth using ChatGPT. A text-to-speech model that produces a human-sounding voice powers the voice feature. But the previous Voice Mode was criticized, particularly after actress Scarlett Johansson hinted that she might sue the corporation for utilizing her voice without permission.
With GPT-4o’s video and audio capabilities, the new Voice Mode should have improved performance and capabilities. For instance, because it relied on three different models, the previous version had latencies of 2.8 seconds (for GPT-3.5) and 5.4 seconds (for GPT-4).
The updated version employs a single model end-to-end for text, vision, and audio, suggesting that a single neural network handles all inputs and outputs. The voice assistant will have four pre-programmed voices, according to the corporation, and the use of the Sky voice type will be discontinued. In addition, the program can now reject requests to make music or other copyrighted content because the corporation has put new filters on it.
With the use of the phone’s camera, the new Voice Mode may provide context for items that users see on their screens. That being said, these functionalities won’t be available in Voice Mode Alpha. Screen and video-sharing features will be released later, the business said.
According to OpenAI, to further refine the model, it will take user input into account. A thorough report on GPT-4o’s performance in August, including safety assessments and constraints, will also be shared.
How to Access Voice Mode Alpha
Only a small number of ChatGPT Plus customers can access the alpha, according to OpenAI’s post. For $20 a month, you can subscribe to ChatGPT Plus. That being stated, you will get instructions via email and a notice in the mobile app if you are chosen for alpha testing. You shouldn’t be concerned if you haven’t received a message because the company will keep adding people on a rolling basis.
In the fall, ChatGPT Plus subscribers will have widespread access to the real-time Voice Mode.
See More: Taco Bell Unveils Drive-Thru With Voice AI, Sets Eyes on Future Tech Innovations