Home How to turn on advanced voice mode in ChatGPT – a guide to new AI feature

How to turn on advanced voice mode in ChatGPT – a guide to new AI feature

TLDR

  • OpenAI has introduced an advanced voice mode for ChatGPT, enabling more natural, real-time conversations with the AI.
  • This feature is available only to Plus and Team users, but not yet in the EU, UK, or specific countries like Iceland, Norway, and Switzerland.
  • Advanced voice uses GPT-4o's native audio capabilities, responding with emotion and real-time adaptability.
  • Users can activate the feature by tapping the voice icon in the app and toggling between standard and advanced voice modes.

Among the many new updates from OpenAI, the tech company behind ChatGPT has announced the rollout of its new advanced voice mode feature, allowing users to have natural conversations with its chatbot.

The company said it’s not yet available in countries in the European Union including Iceland, Liechtenstein, Norway, Switzerland or the United Kingdom.

OpenAI’s co-founder and CEO, Sam Altman, wrote in a post on X: “Hope you think it was worth the wait.”

Here’s what you need to know about it and how to turn advanced voice mode on in ChatGPT.

What is advanced voice mode on ChatGPT?

Voice conversations allow users to have a spoken conversation with ChatGPT, which means you can have more natural interactions and conversations. When you can ask questions or have discussions through voice input, ChatGPT will give a spoken response.

There are currently two types of voice conversations – standard and advanced.

ReadWrite reported on OpenAI launching its new standard voice mode last month. Standard voice uses several large language models (LLMs) to generate its response, including transcribing what you say into text before sending it to OpenAI’s models for response. While standard voice is not generally multimodal like advanced voice, standard voice conversations also use GPT-4o alongside GPT-4o mini. Each prompt in standard voice counts towards your message limits.

Where the advanced mode differs is that it uses GPT-4o’s native audio capabilities and features. As a result, OpenAI hopes to produce more natural, real-time conversations that pick up on non-verbal cues, such as the speed the user is talking and can respond with emotion.

However, usage of advanced voice by Plus and Team users is limited daily.

How to activate voice mode in ChatGPT?

In July, OpenAI introduced an audio-only Advanced Voice Mode to a small group of ChatGPT Plus users, with plans to expand it to all subscribers this fall.

While screen and video sharing were part of the initial demo, they are currently not available in this alpha release, and OpenAI has not provided a timeline for their inclusion.

Plus subscribers will receive an email notification when the feature is available to them. Once activated, users can toggle between Standard and Advanced Voice Modes at the top of the app when using ChatGPT’s voice feature.

To start a voice conversation, tap the Voice icon in the bottom-right corner of your screen.

If you’re using advanced voice, you’ll see a blue orb in the center of the screen when the conversation begins. For standard voice, the orb will be black instead.

A smartphone screen displaying a notification about nearing the daily limit for advanced voice usage. The message reads, "Almost at daily limit. You have 15 minutes remaining of advanced voice. You can switch to standard voice when you hit the limit. Learn more." A circular blue and white gradient is shown in the middle of the screen, with microphone and settings icons visible at the top.
OpenAI’s new advanced voice mode for ChatGPT shows a blue orb. Credit: OpenAI

During the conversation, you can mute or unmute yourself by tapping the microphone icon at the bottom left. And when you’re ready to end the chat, just hit the exit icon in the bottom right.

If it’s your first time starting a voice chat, or the first time using advanced voice, you’ll be asked to pick a voice. Just a heads-up, the volume in the selector might be a bit different from what you hear in the conversation.

A smartphone screen introduces advanced voice mode with the heading "Say hello to advanced voice mode." Key features include "Natural conversations" that respond to interruptions and humor, "Multiple voices" with more voice options, "Personalized to you" using memory and custom instructions, and "You're in control" allowing audio recordings to be saved or deleted. At the bottom, a "Continue" button is displayed.
The advanced voice mode feature is being rolled out to some Plus users. Credit: Suswati Basu for ReadWrite

You can always change your voice in the settings later, and advanced voice users can even adjust their voice directly from the conversation screen using the customization menu in the top-right.

Make sure you’ve given the ChatGPT app permission to use your microphone so everything works smoothly.

And if this feature isn’t available for you yet, you’ll see a headphones icon instead of the mute/unmute buttons. With both versions, you can interrupt the conversation, steering it in a way that feels more appropriate for you.

Is ChatGPT voice available?

If you’re signed in to ChatGPT through the iOS, macOS, or Android apps, you already have access to the standard voice feature. However, advanced voice is currently only available to Plus and Team users.

There’s a daily limit for using advanced voice, which might change over time, but you’ll get a heads-up when you’re close to the limit—starting with a 15-minute warning. Once you hit the limit, your conversation will switch to standard voice automatically.

The advanced voice doesn’t support things like images yet, so users can only continue an advanced voice conversation with text or standard voice, not vice versa. Conversations started in standard voice can always be resumed using standard voice or text, but not advanced voice. Advanced voice isn’t available with GPTs either—you’ll have to switch to standard voice for that.

OpenAI hasn’t introduced certain accessible features either. Consequently, subtitles aren’t available during voice conversations, but the transcription will appear in your text chat afterward. Also, you can only have one voice chat at a time.

Advanced voice can create and access memories as well as custom instructions, just like standard voice, which also has these features.

Is ChatGPT voice chat safe?

In August, OpenAI revealed there were some security flaws with ChatGPT’s voice mode, but reassured that they were on top of it. OpenAI published a report on GPT-4o’s safety features, addressing known issues that occur when using the model.

The “safety challenges” with ChatGPT’s voice mode include typical concerns like generating inappropriate responses, such as erotic or violent content and making biased assumptions. OpenAI has trained the model to block such outputs, but the report highlights that nonverbal sounds, like erotic moans, violent screams, and gunshots, aren’t fully filtered. This means prompts involving these sensitive sounds might still trigger responses.

Another challenge is communicating with the model vocally. Testers found that GPT-4o could be tricked into copying someone’s voice or accidentally sounding like the user. To prevent this, OpenAI only allows pre-approved voices – not including a Scarlett Johansson-like voice, which the company has already removed. In addition, while GPT-4o can recognize other voices, it has been trained to reject such requests for privacy reasons, unless it’s identifying a famous quote.

Red-teamers also flagged that GPT-4o could be manipulated to speak persuasively, which poses a bigger risk when spreading misinformation or conspiracy theories, given the impact of spoken words. The model has been trained to refuse requests for copyrighted content and has extra filters to block music. And as a fun fact, it’s programmed not to sing at all. However, in this example from a user on X, the voice helps them to tune his guitar by humming the note.

How can I stop sharing audio?

You can stop sharing your audio anytime by going to the data controls page in your ChatGPT settings. Just toggle off the “Improve voice for everyone” setting.

If you don’t see “Improve voice for everyone” in your Data Controls settings, that means you haven’t shared your audio with OpenAI, and it’s not being used to train models.

If you choose to stop sharing, audio from future voice chats won’t be used for model training. However, audio clips that were previously disassociated from your account may still be used to train OpenAI’s models.

OpenAI also mentioned that even if you stop sharing audio, it “may still use transcriptions of those chats to train our model” if the “Improve the model for everyone” setting is still on. To fully opt-out, disable “Improve the model for everyone.”

Audio clips from your advanced voice chats will be stored as long as the chat remains in your chat history. If you delete the chat, the audio clips will also be deleted within 30 days, unless they’re needed for security or legal reasons. If you’ve shared your audio clips with OpenAI to help train models, those clips may still be used, but only after they’ve been disassociated from your account.

Featured image: Ideogram / Canva

About ReadWrite’s Editorial Process

The ReadWrite Editorial policy involves closely monitoring the tech, gambling and blockchain industries for major developments, new product and brand launches, AI breakthroughs, game releases and other newsworthy events. Editors assign relevant stories to in-house staff writers with expertise in each particular topic area. Before publication, articles go through a rigorous round of editing for accuracy, clarity, and to ensure adherence to ReadWrite's style guidelines.

Suswati Basu
Tech journalist

Suswati Basu is a multilingual, award-winning editor and the founder of the intersectional literature channel, How To Be Books. She was shortlisted for the Guardian Mary Stott Prize and longlisted for the Guardian International Development Journalism Award. With 18 years of experience in the media industry, Suswati has held significant roles such as head of audience and deputy editor for NationalWorld news, digital editor for Channel 4 News and ITV News. She has also contributed to the Guardian and received training at the BBC As an audience, trends, and SEO specialist, she has participated in panel events alongside Google. Her…

Get the biggest tech headlines of the day delivered to your inbox

    By signing up, you agree to our Terms and Privacy Policy. Unsubscribe anytime.

    Tech News

    Explore the latest in tech with our Tech News. We cut through the noise for concise, relevant updates, keeping you informed about the rapidly evolving tech landscape with curated content that separates signal from noise.

    In-Depth Tech Stories

    Explore tech impact in In-Depth Stories. Narrative data journalism offers comprehensive analyses, revealing stories behind data. Understand industry trends for a deeper perspective on tech's intricate relationships with society.

    Expert Reviews

    Empower decisions with Expert Reviews, merging industry expertise and insightful analysis. Delve into tech intricacies, get the best deals, and stay ahead with our trustworthy guide to navigating the ever-changing tech market.