Skip to main content

ChatGPT Advanced Voice mode: release date, compatibility, and more

Advanced Voice Mode is a new feature for ChatGPT that enables users to hold real-time, humanlike conversations with the AI chatbot without the need for a text-based prompt window or back-and-forth audio. It was released in late July to select Plus subscribers after being first demoed at OpenAI’s Spring Update event.

According to the company, the feature “offers more natural, real-time conversations, allows you to interrupt at any time, and senses and responds to your emotions.” It can even take breath breaks and simulate human laughter during conversation. The best part is that access is coming soon, if you don’t have it already.

When will I get Advanced Mode?

Introducing GPT-4o

Last week, OpenAI finally released its highly anticipated Advanced Voice feature to a select few of its ChatGPT-Plus subscribers. So, some people are already out there using it. We don’t know how large this initial rollout is, however, or how quickly OpenAI is inviting new members.

The company promises to roll it out to more users in the coming weeks, and to all Plus subscribers by this fall, although many users remain impatient to try it for themselves. So, the simple answer is that you’ll get access to it sometime between now and the fall, barring any further delays. “Fall” can be vague, but if we don’t see a full rollout to all users by December, that would constitute a delay.

Either way, you’ll know you have access when you receive an email inviting you to give it a try. You’ll also see a notification when you open the ChatGPT app.

What do I need to run Advanced Voice Mode on my phone?

the advanced voice alert on the chatgpt app
OpenAI

In addition to a Plus subscription, users will need an Android handset running app version 1.2024.206 or later, or an iPhone using iOS 16.4 or later and app version 1.2024.206 or later. Unfortunately, just having the right equipment isn’t enough to guarantee you a spot in the alpha release phase. What’s more, OpenAI has not released any details as to how or why it chooses the users it does.

If you do win the alpha release lottery and are selected, OpenAI will send you an email about it. You’ll also see a tooltip in the bottom-right corner of the ChatGPT mobile app that allows you to select the new feature.

Is OpenAI going to use my conversations to train its model?

Of course. Throughout the alpha release phase, OpenAI plans to leverage audio from conversations with advanced Voice Mode to train its models, assuming you haven’t yet turned off the app’s dat- sharing option.

Doing so is actually quite simple. On your mobile app, go to the Data Controls tab in your Settings menu and deselect Improve voice for everyone.

How often and how much can I use Advanced Voice Mode?

10 Minutes with ChatGPT's Advanced Voice Mode

According to OpenAI, both the inputs and outputs for Advanced Voice come with daily usage limits, however, there’s no specifics on exactly how long those are and “precise limits are subject to change.” That said, user Himels Tech has already posted a video of themselves conversing with the AI for the better part of 10 minutes.

The AI will prompt users when they have 3 minutes of chatting left, before ending the conversation and sending the user back to the standard voice interface.

What can I, and can’t I, do with Advanced Voice Mode?

Khan!!!!!! pic.twitter.com/xQ8NdEojSX

— Ethan Sutin (@EthanSutin) July 30, 2024

Advanced Voice Mode is, at its core, simply a new way to interact with the same GPT-4o large language model that people already use for their text-based queries. So, in short, most anything you can do with ChatGPT, you can do with Advanced Voice, but with funny voices. From beatboxing to storytelling to counting really, really fast, early users are already putting the new feature through its paces.

There are safety guardrails and feature limits to what users can ask of the new mode, however. For one, users can’t use Advanced Voice to make new memories, nor can they use custom instructions or access GPTs using it. And while the AI will remember previous Advanced Voice conversations and be able to recall details of those talks, it cannot yet access previous chats conducted through the text prompt or the standard voice mode.

What’s more, Advanced Voice will not sing, no matter how you ask. Per the company, “to respect creators’ rights, we’ve put in place several mitigations, including new filters, to prevent advanced Voice Mode from responding with musical content including singing.”

Andrew Tarantola
Andrew has spent more than a decade reporting on emerging technologies ranging from robotics and machine learning to space…
All the wild things people are doing with ChatGPT’s new Voice Mode
Nothing Phone 2a and ChatGPT voice mode.

ChatGPT's Advanced Voice Mode arrived on Tuesday for a select few OpenAI subscribers chosen to be part of the highly anticipated feature's alpha release.

The feature was first announced back in May. It is designed to do away with the conventional text-based context window and instead converse using natural, spoken words, delivered in a lifelike manner. It works in a variety of regional accents and languages. According to OpenAI, Advanced Voice, "offers more natural, real-time conversations, allows you to interrupt anytime, and senses and responds to your emotions."

Read more
ChatGPT’s highly anticipated Advanced Voice could arrive ‘next week’
screencap. two people sitting at a desk talking to OpenAI's Advanced Voice mode on a cellphone

OpenAI CEO and co-founder Sam Altman revealed on X (formerly Twitter) Thursday that its Advanced Voice feature will begin rolling out "next week," though only for a few select ChatGPT-Plus subscribers.

The company plans to "start the alpha with a small group of users to gather feedback and expand based on what we learn."

Read more
GPT-4: everything you need to know about ChatGPT’s standard AI model
A laptop opened to the ChatGPT website.

People were in awe when ChatGPT came out, impressed by its natural language abilities as an AI chatbot originally powered by the GPT-3.5 large language model. But when the highly anticipated GPT-4 large language model came out, it blew the lid off what we thought was possible with AI, with some calling it the early glimpses of AGI (artificial general intelligence).
What is GPT-4?
GPT-4 is the newest language model created by OpenAI that can generate text that is similar to human speech. It advances the technology used by ChatGPT, which was previously based on GPT-3.5 but has since been updated. GPT is the acronym for Generative Pre-trained Transformer, a deep learning technology that uses artificial neural networks to write like a human.

According to OpenAI, this next-generation language model is more advanced than ChatGPT in three key areas: creativity, visual input, and longer context. In terms of creativity, OpenAI says GPT-4 is much better at both creating and collaborating with users on creative projects. Examples of these include music, screenplays, technical writing, and even "learning a user's writing style."

Read more