Curated by THEOUTPOST
On Wed, 25 Sept, 4:04 PM UTC
14 Sources
[1]
ChatGPT Advanced Voice Conversation Skills Tested
Open AI recently showcased its ChatGPT advanced voice capabilities in a live stream, where the host interacted with the AI in real-time. This in-depth demonstration provided valuable insights into the current state and potential of AI voice technology. Let's dive into a comprehensive evaluation of Open AI's voice AI, examining its strengths, limitations, and future prospects. ChatGPT Advanced voice is available in the iOS / Android mobile apps as of version 1.2024.261 or later and is rolling out to all Plus and Team users in the ChatGPT app over the course of the week. Unfortunately ChatGPT Advanced Voice is not yet available in the EU, the UK, Switzerland, Iceland, Norway, and Liechtenstein. But let us hope that OpenAI rolls it out very soon. One of the most impressive aspects of Open AI's voice AI is its ability to engage in dynamic, real-time conversations. During the live stream, the AI consistently responded promptly and accurately to a wide range of prompts and questions posed by the host. This seamless interaction showcased the AI's potential for natural, flowing communication, making it well-suited for applications such as: The AI's responsiveness and contextual understanding demonstrate significant advancements in natural language processing and machine learning algorithms. As this technology continues to evolve, we can expect even more sophisticated and human-like conversational abilities from AI systems. Here are a selection of other articles from our extensive library of content you may find of interest on the subject of ChatGPT Voice: Another standout feature of Open AI's voice AI is its impressive proficiency in multiple languages and accents. Throughout the live stream, the AI successfully understood and responded to prompts in various languages, showcasing its potential for global applications. This multilingual capability is a significant step towards making AI more accessible and useful for people worldwide. Moreover, the AI demonstrated a remarkable ability to recognize and adapt to different accents within a language. This flexibility enhances its usability in diverse linguistic contexts and ensures a more inclusive user experience. As AI continues to break down language barriers, it opens up exciting possibilities for cross-cultural communication and collaboration. Open AI's voice AI showcased its versatility through various role-playing scenarios during the live stream. The host tested the AI's performance in simulating roles such as tech support, news reporting, and storytelling. The AI adapted to these scenarios with ease, providing relevant and contextually appropriate responses. This ability to simulate different roles highlights the AI's potential for: As AI continues to refine its role-playing capabilities, it can transform how we learn, work, and engage with technology in our daily lives. One intriguing aspect explored during the live stream was the AI's ability to recognize and respond to different emotional tones in the host's voice. While the AI showed some capability in detecting emotions like happiness, sadness, and anger, there were instances where its responses did not accurately reflect the intended emotional tone. This highlights that emotion recognition in AI is still a developing frontier. As research progresses in this area, we can expect more refined and reliable performance in detecting and responding to human emotions. Advancements in emotional intelligence will be crucial for creating AI systems that can provide empathetic and emotionally appropriate interactions. Despite its impressive capabilities, Open AI's voice AI is not without limitations. During the live stream, the AI encountered challenges with certain tasks, such as singing and specific character impersonations. These limitations serve as reminders of the current boundaries of AI voice technology and indicate areas that require further development and refinement. Additionally, the AI's usage was subject to rate limits of approximately 70 minutes per day. While these restrictions help manage computational resources and ensure optimal performance, they may impact the AI's potential for continuous, long-duration interactions. As AI infrastructure and algorithms advance, we can anticipate more efficient resource management and expanded usage capabilities. Looking ahead, the live stream also sparked discussions about future updates to Open AI's voice AI, including the potential integration of vision capabilities. Combining voice and vision could significantly enhance the AI's understanding and interaction with the world, allowing it to process and respond to both auditory and visual cues. This development holds immense promise for creating more comprehensive and immersive AI experiences. Open AI's advanced voice update demonstrates the rapid progress and exciting potential of AI voice technology. With its real-time engagement, multilingual proficiency, and scenario simulation capabilities, this AI system offers a glimpse into the future of human-machine interaction. As researchers and developers continue to push the boundaries of AI, we can anticipate even more sophisticated and intuitive voice-based AI systems that will transform various aspects of our lives.
[2]
OpenAI Finally Rolls Out Advanced Voice to ChatGPT Plus and Team Users
After much wait, OpenAI has finally begun rolling out its highly anticipated Advanced Voice feature to all Plus and Team users in the ChatGPT app. The rollout is expected to be completed over the course of the week, with notifications appearing in the app once access is granted. The new feature introduces five distinct voices along with support for over 50 languages, allowing users to hear responses in different accents. Advanced Voice now also includes the ability to customise instructions, a feature that allows for greater personalisation. Users can set specific preferences for how the AI interacts via voice, enhancing the overall experience. While Advanced Voice offers a range of improvements, it is currently unavailable in the European Union, the UK, Switzerland, Iceland, Norway, and Liechtenstein. This rollout follows several key updates to the ChatGPT app, including the addition of Custom Instructions, Memory, and enhanced voice accents. OpenAI released GPT-4o at its latest Spring Update event earlier this year, which won hearts with its 'omni' capabilities across text, vision, and audio. OpenAI's demos, which included a real-time translator, a coding assistant, an AI tutor, a friendly companion, a poet, and a singer, soon became the talk of the town. However, its Advanced Voice Mode wasn't released. Recently, Kyutai, a French non-profit AI research laboratory, launched Moshi, a real-time native multimodal foundational AI model capable of conversing with humans in real time, much like what OpenAI's advanced model was intended to do. Hume AI introduced EVI 2, a new foundational voice-to-voice AI model that promises to enhance human-like interactions. Available in beta, EVI 2 can engage in rapid, fluent conversations with users, interpreting tone and adapting its responses accordingly. The model supports a variety of personalities, accents, and speaking styles and includes multilingual capabilities. Meanwhile, Amazon Alexa is partnering with Anthropic to improve its conversational abilities, making interactions more natural and human-like. Earlier this year, Google launched Astra, an 'universal AI agent' built on the Gemini family of AI models. Astra features multimodal processing, enabling it to understand and respond to text, audio, video, and visual inputs simultaneously.
[3]
OpenAI's controversial ChatGPT advanced voice mode feature finally available to users: All you need to know about it
The much awaited OpenAI 'advanced audio' chat is all set to come to your ChatGPT The much awaited OpenAI 'advanced audio' chat is all set to come to your ChatGPT. On Tuesday OpenAI said that it's most loved chatbot can now give you a 'more natural' conversation. With the new update you can experience smooth conversations, just like we do with humans. On September 24, OpenAI wrote in an official blog that its 'advanced voice' feature will roll out to ChatGPT users . As of now the feature is only available to ChatGPT plus users. So, here's how the new feature is going to change your 'conversation with AI' experience. OpenAI rolls 'advanced voice' mode for ChatGPT According to OpenAI's official blog, the feature will continue to roll out during this entire week. Notably, the feature is not yet available in EU countries, Switzerland, Liechtenstein, Iceland, Norway, or the U.K, the company highlighted. Early news showed that this 'human-like' voice feature of ChatGPT had initially faced a backlash. Popular Hollywood actors and actresses had alleged it of copying their voice without their consent. This had delayed the launch from May end to July end. After all the fighting, it's finally here! How cool can it be if artificial intelligence (AI) could talk with you? Yes, that's what OpenAI is aiming for with its 'advanced voice' mode. With the new feature you get five distinct voices. Additionally, the voices come with support for over 50 languages. Now, that's a long list of languages! You get to interact with ChatGPT with any language you want. Please note that to get these features, you need to update your existing ChatGPT. What's next The 'advanced voice' feature will also allow you to hear responses in different accents. In addition to this the 'advanced voice' also includes the ability to customize instructions. Moreover you get the scope for greater personalisation. For example you can set specific preferences for how the AI interacts through voice. Initially the rollout got plenty of publicity because of a voice named ' Sky'. This voice sounded similar to Scarlett Johansson in the 2013 movie "Her." This raised a lot of chaos and the actress filed a legal case. So, after a long delay and a certain legal process, the feature finally got rolled out. To keep up with the excitement, Sam Altman showed a cheering approach with a social media post. On Tuesday "Hope you think it was worth the wait," Sam Altman, OpenAI's CEO and co-founder, wrote in an X (earlier Twitter) post.
[4]
OpenAI Finally Brings Advanced Audio Chats to ChatGPT After Months-Long Pause
Samantha Kelly is a freelance writer with a focus on consumer technology, AI, social media, Big Tech, emerging trends and how they impact our everyday lives. Her work has been featured on CNN, NBC, NPR, the BBC, Mashable and more. OpenAI is rolling out an advanced audio mode for its paid ChatGPT users several months after hitting pause on the feature. The update brings five new voices (for a total of nine), improved accents, improved memory and more personalized interactions to the audio tool available for ChatGPT Plus and Team subscribers. Plans start at $20 a month. The feature, originally announced in May, made headlines after one of the voice options sounded eerily similar to the fictional voice assistant portrayed by actress Scarlett Johansson in the film "Her." After Johansson's legal team sent letters to OpenAI, the company delayed the greater rollout of advanced audio mode. In a blog post at the time, OpenAI said the specific voice, which the company called Sky, "was never intended to resemble (Johansson's)." "Out of respect for Ms. Johansson, we have paused using Sky's voice in our products," the post said. "We are sorry to Ms. Johansson that we didn't communicate better." The Sky voice option was removed from the feature. The update promises the ability to better recall conversations and have more natural interactions. It also has the ability to converse in 50 languages and will listen to requests if interrupted. "Hope you think it was worth the wait," CEO Sam Altman wrote in a tweet. Advanced audio mode is not yet available in all regions, including the EU, UK, Switzerland, Iceland, Norway, and Liechtenstein. The launch comes as Google continues to roll out its free AI-powered Gemini Live audio tool to English-language Android users through the Gemini app. Gemini Live, which promises to learn and adapt to users' conversational style over time, was also announced in May. Meanwhile, Meta announced on Wednesday it is adding celebrity voices, including those of Kristen Bell, John Cena and Dame Judi Dench, to its AI chatbot.
[5]
ChatGPT's Advanced Voice Feature Is Rolling Out to More Users
If you have ever wanted to have a full-blown conversation with ChatGPT, now you can. That is, as long as you pay for the privilege of using ChatGPT. More paid users are gaining access to ChatGPT's Advanced Voice Mode (AVM), which is designed to make interacting with ChatGPT feel more natural. Advanced Voice Comes to ChatGPT Plus and Teams OpenAI has started rolling out ChatGPT's new Advanced Voice Mode to more users. The feature will initially be made available to people on the ChatGPT Plus and ChatGPT Teams tiers, with Enterprise and Edu users not far behind. While ChatGPT's Advanced Voice Mode is going to be accessible in most countries, there are some notable exceptions. AVM is not currently available in the EU, the UK, Switzerland, Iceland, Norway, or Liechtenstein. This is likely due to more stringent regulations surrounding AI in these territories. OpenAI Has Made Improvements to Advanced Voice OpenAI has made some improvements to AVM since it initially showcased the technology in May. For starters, AVM has been redesigned, with a blue sphere representing the chatbot rather than the series of black dots originally used. There are also five new voices to try, taking the total number to nine. The new voices are named Arbor, Maple, Sol, Spruce, and Vale, and join the already-available Breeze, Juniper, Cove, and Ember. All of these voices have been named after natural elements, which plays into the idea that AVM makes ChatGPT more natural. One voice no longer available is Sky, which Scarlett Johansson, who famously played an AI in Her, claimed sounded too similar to her own. OpenAI also claims to have improved ChatGPT's AVM in several notable (and noticeable) ways. These include conversational speed, smoothness, and accents in foreign languages. The company has also added Custom Instructions, which bring some of the customization options from the text version to the voice version. Essentially, you can customize ChatGPT to respond to you in a certain way. Lastly, Memory allows ChatGPT to remember voice conversations it has had with you, allowing you both to reference them later on. Voice Interactions Makes AI Feel Exciting Again I have to admit that being able to speak to an AI, and have it reply in kind, is much more desirable than messaging back and forth. Not only does it feel more natural, it makes artificial intelligence feel less artificial and more real. Just as Alexa did when it first came out. I remember the excitement of trying ChatGPT for the first time, and thinking about how AI could change the world (for better or worse). But that early enthusiasm has since died off somewhat. Maybe ChatGPT's Advanced Voice Mode will trigger a new wave of excitement.
[6]
Know Details of OpenAI's New Voice Assistant for ChatGPT Users
OpenAI has released a pioneering feature of voice assistant available to all paid customers. Through this new voice mode, humans will interact more naturally and effectively with AI. Here is a comprehensive look at how this feature is bound to transform the way we interact with AI. First off, it already has been noted that the is now promising a new style with multiple voices. So far, five new voices have been added, and there are nine total voices in total. All these voices can be personalized to suit the user's preference for better enjoyment of the conversation. Whether you wish for a smooth calm soothing voice or an energetic one, there is a voice for everyone. The new voices have improvements in accent which sounds more natural and realistic. That is going to be helpful for the variety of different linguistic backgrounds. It's striving to carry it over to a more natural, real-time conversation with a lot more flow, thereby making an interaction with ChatGPT similar to a chat with a human. Probably the biggest leap that's made in a new voice assistant is its ability to note and interpret how emotionally the person speaks. This means that it would know if the person is happy, sad, or frustrated. So, the AI has a sense of how to respond with something that would be apt for the user's situation. For instance, if a user seems to be upset, it may provide comforting words or solutions. Bringing a connection much more empathetic and supportive during an interaction. It represents significant strides toward meaningful and engaging conversations.
[7]
ChatGPT's Advanced Voice is rolling out to subscribers | Digital Trends
OpenAI announced via Twitter on Tuesday that it will begin rolling out its Advanced Voice feature, as well as five new voices for the conversational AI, to subscribers of the Plus and Teams tiers throughout this week. Enterprise and Edu subscribers will gain access starting next week. Recommended Videos Advanced Voice, which runs on the GPT-4o model, allows users to forgo written text prompts and speak directly with the chatbot as they would another person. It was first announced at OpenAI's Spring Update event and released to a select group of ChatGPT Plus subscribers to beta test the system in July. Now, every paying subscriber will be able to try the feature for themselves. The company also unveiled five new voices for the chatbot: Arbor, Maple, Sol, Spruce, and Vale (you can listen to them right now). They'll be available in both Standard and Advanced Voice modes, joining the four voices -- Breeze, Juniper, Cove, and Ember -- that ChatGPT already offers. OpenAI also noted that, while video and screen sharing are not currently supported in Advanced Voice, those capabilities will roll out at a later date. What's more, OpenAI is incorporating a pair of tools to grant Advanced Voice capabilities more in line with the rest of the text-based chatbot experience: memory and custom instructions. When it first debuted, Advanced Voice could only reference information from the chat it was currently having. With the memory function, the AI will be able to recall details from previous conversations as well, reducing the need for users to repeat themselves. Similarly, custom instructions are designed to set ground rules for the model to follow when generating its responses. For example, you could dictate that any coding-based responses be presented in Python. Plus and Teams subscribers will receive an in-app notification when the feature goes live on their account. Unfortunately, Advanced Voice is not available in the EU, the U.K., Switzerland, Iceland, Norway, and Liechtenstein. ChatGPT isn't the only AI that can converse directly with its users. Tuesday's announcement comes less than a fortnight after Google released its Gemini Live to all users -- including those on the free tier.
[8]
Massive ChatGPT Voice Update Makes AI Conversation Even More Lifelike
ChatGPT, the popular AI-powered conversational tool, has recently undergone a significant voice upgrade that promises to enhance user interactions with advanced AI Voice features. This update includes custom instructions, memory retention, multiple voices, and improved accents, all designed to make conversations more engaging and lifelike. While these enhancements aim to transform the way users interact with the AI, it's important to note that the core functionality of ChatGPT remains largely unchanged. Sam Altman, the CEO of OpenAI, announced that the advanced voice mode would be available to users within a week of the initial announcement. This new mode introduces several features that are designed to personalize and enrich user interactions with ChatGPT. Some of the key features include: The user experience with the new voice mode has been mixed so far. Some users have successfully accessed the new features by reinstalling the app, while others have faced challenges in getting the upgrade to work properly. Demonstrations of the voice mode's capabilities highlight its potential for emotional storytelling and accent variations, making conversations more dynamic and engaging. However, it's important to keep in mind that while the new features offer novelty and entertainment value, they do not significantly alter the core functionality of ChatGPT. The AI's ability to understand and respond to user queries remains largely the same, with the added benefit of a more personalized and engaging conversational experience. Two types of ChatGPT Voice conversations, will be available Standard and Advanced. Here are a selection of other articles from our extensive library of content you may find of interest on the subject of ChatGPT voice : One of the most notable aspects of the voice upgrade is the enhanced conversational experience it offers. The ability to switch between different voices and accents adds a layer of personalization and fun to interactions with ChatGPT. Users can now engage in more dynamic and entertaining conversations, with the AI adapting to their preferences and communication style. However, it's crucial to understand that while the upgrade enhances the user experience, it does not fundamentally change the practical utility of ChatGPT. The AI's ability to provide information, answer questions, and assist with tasks remains similar to previous versions. The new features primarily focus on improving the engagement and entertainment value of conversations rather than expanding the AI's core capabilities. ChatGPT's voice upgrade is part of the broader advancements in AI technology that are currently taking place. Generative AI, which focuses on creating new content based on patterns and rules learned from existing data, continues to evolve and impact various sectors, including healthcare, finance, and autonomous vehicles. AI agents like ChatGPT are becoming more sophisticated, with large language models improving their ability to understand and generate human-like text. These advancements contribute to the overall enhancement of AI capabilities, making tools like ChatGPT more versatile and effective in engaging with users and providing valuable assistance. The future prospects for ChatGPT and similar AI technologies are promising. As AI continues to evolve and improve, we can expect to see more advanced features and capabilities that enhance user experience and expand the practical applications of these tools. The ability to engage in more human-like conversations could increase user interaction and satisfaction, making AI-powered tools more appealing and accessible to a wider audience. There is also speculation about the potential for emotional attachment to AI, as interactions become more personalized and lifelike. Integration with other technologies, such as augmented reality devices like Meta Ray-Ban glasses, could further expand the utility and appeal of AI-driven tools. As ChatGPT and other AI technologies continue to advance, we can expect to see new and innovative ways in which they can be applied to various aspects of our lives. For users considering upgrading to the pro version of ChatGPT, the new voice features may be a significant factor in their decision-making process. The enhanced conversational experience and the novelty of the advanced voice mode can greatly contribute to personal enjoyment and engagement with the AI. While the practical utility of ChatGPT remains largely unchanged, the improved user experience and potential for future integrations may make the upgrade worthwhile for those seeking a more engaging and personalized interaction with AI technology. ChatGPT's new voice upgrade represents a significant step forward in the development of conversational AI technology. While the core functionality remains largely unchanged, the advanced features and enhanced user experience offer a glimpse into the future potential of AI-powered tools. As ChatGPT and other AI technologies continue to evolve, we can expect to see more innovative and engaging ways in which they can be applied to various aspects of our lives, transforming the way we interact with technology and each other.
[9]
OpenAI Rolls Out Advanced Voice Mode to All Plus and Team Subscribers
OpenAI is expanding access to the Advanced Voice mode it first showed off in May. "Advanced Voice is rolling out to all Plus and Team users in the ChatGPT app over the course of the week," according to a tweet from OpenAI, which clarified later that the rollout is not happening in the EU, the UK, Switzerland, Iceland, Norway, and Liechtenstein yet, likely due to regulations there. The feature launched for select Plus subscribers in July, and promises to provide a natural-sounding back and forth with the AI. In a May demo with OpenAI employees, it easily understood what they asked it, and responded in real time with accurate, useful information, even if it was interrupted or the employee adjusted their query mid-sentence. Since then, OpenAI has added "Custom Instructions, Memory, five new voices, and improved accents." There were also improvements made to conversational speed. The new voices include two male and three female voices. To show off the improved accents, one of the female voices has an English accent and one of the male voices is Australian. The new voices increase the total number available to nine. Still missing is Sky, which drew the ire of Scarlett Johansson earlier this year and has since been disabled. In another tweet, OpenAI posted a video to show off custom instructions and memory. For custom instructions, there will be a spot in the settings where people can tell ChatGPT how they want the bot to talk to them. As an example, someone may want the bot to refer to them by a nickname or speak more slowly or more clearly. It'll also remember conversations and instructions you gave it so that it can reference those things again later. So, if you tell ChatGPT that you're from a specific town, it'll suggest stuff to do around town when you ask for recommendations later. Team and Plus users will get a notification when it's their turn to access advanced voice mode.
[10]
OpenAI Launches Advanced Voice Mode, Minus the Scarlett Johansson Drama - Decrypt
OpenAI has begun rolling out its much-anticipated Advanced Voice Mode for ChatGPT Plus and Teams users, marking another step towards a more human-like AI interaction. The feature allows for real-time, fluid conversations powered by GPT-4o, OpenAI's latest model, which combines text, vision, and audio to deliver faster responses. "Advanced Voice is rolling out to all Plus and Team users in the ChatGPT app over the course of the week," OpenAI said in an official tweet, "It can also say "Sorry I'm late" in over 50 languages," it added -- addressing the long delay this project went through. Needless to say, one notable element is still missing: the flirty and definitely too human-like "Sky" voice, which caused a stir for its uncanny resemblance to actress Scarlett Johansson. After her legal team sent letters to OpenAI's CEO Sam Altman, OpenAI put the Sky voice on hold, maintaining that any resemblance between Johansson's distinctive voice and Sky was purely coincidental. Instead, OpenAI introduced five new voices: Arbor, Maple, Sol, Spruce, and Vale, which are available in both Standard and Advanced Voice Mode. These join the previously available Breeze, Juniper, Cove, and Ember. (For some reason, the company seems to be naming them after soap fragrances.) Users in the Plus and Team tiers will gradually gain access to these new voices, designed to make conversations more natural, with emotional responsiveness and the ability to interrupt and switch topics on the fly. Additionally, OpenAI is adding compatibility with custom instructions and "memories" to allow users to personalize their ChatGPT experience further, tailoring interactions to their preferences. Just as the text-based chatbot learns from your instructions (i.e., your name, occupation, and probably the type of answers you like to read), the new voices will try to learn from your conversations, making them more natural, familiar, and used to your preferences. Users in the EU, UK, Switzerland, Iceland, Norway, and Liechtenstein will have to wait, as the feature has not yet rolled out in those regions. Enterprise and Edu users can expect access starting next week, according to OpenAI's timeline. The rollout is slow, and not all users, even from supported regions, have the feature available. OpenAI also refined accents in popular foreign languages and enhanced conversational speed and smoothness. The design has also been updated, with an animated blue sphere that visually represents the voice interaction as it happens and is more aesthetically pleasing than the minimalist black dot they used to show. While OpenAI continues to refine its voice AI offerings, competition in the space has been heating up. Google's NotebookLM currently sets the bar with some of the most human-like AI voices available, able to simulate entire debates between AI-generated speakers with remarkable realism. Google's AI tool can process up to one million data tokens and let users interact with it, Decrypt previously reported. Once users upload a specific group of documents with different types of information, Notebook LM can generate up to 10 minutes of audio with two AIs talking about that specific information. The result is almost extremely realistic. Besides Google, Meta has also entered the fray with its own live assistant, Meta AI, though it is not yet widely available. The assistant is also capable of having natural conversations with users, processing commands fluently. The voice is more natural than the typically robotic voice we see in most AI assistants, but it still has some giveaways -- like the speech cadence and speed -- that make it identifiable as AI-generated. That said, Reuters has reported that Meta's upcoming chatbot will have the personas of Judy Dench and Michael Cerna. It's not Scarlet Johansson, but nor is it chopped liver.
[11]
OpenAI rolls out more realistic Advanced Voice Mode for premium users
OpenAI has started rolling out the advanced voice mode for premium users of its AI chatbot ChatGPT. The audio feature is touted to be more natural and immersive and will be released through the week. The AI firm has said that it's not yet available in EU countries including Iceland, Norway, Switzerland, Liechtenstein or the U.K. After ChatGPT Plus and Teams customers, the feature will be released for Enterprise and Edu users from next week. CEO Sam Altman announced the feature on X saying, "Hope you think it was worth the wait." ChatGPT will also be getting five new voices that users can test called Arbor, Maple, Sol, Spruce and Vale, making the total number of voice options nine. OpenAI's stunning $150 billion valuation hinges on upending corporate structure While OpenAI had announced the advanced voice feature in May, the rollout was pushed back after actress Scarlett Johansson alleged that the demo voice called Sky was distinctly similar to her own in the 2013 film, "Her." While the company responded to the claim saying they hadn't modelled the voice after Johansson's but paused using it after her legal counsel sent letters stating they didn't have permission. For premium subscribers, they need to open the ChatGPT app to receive a notification once they have turned on access to the new feature. Then, the user can click on 'Continue' to go on. Once they start a new chat, a sound wave icon will appear next to the microphone icon and the 'Message' field. A small bump sound will be heard after a few seconds post which the circle in the middle of the screen that will then transform into a sky blue and white animation. The bot responds then. Users can choose a different accent or speed up a conversation. Published - September 25, 2024 11:20 am IST Read Comments
[12]
ChatGPT Advanced Voices Are Rolling out to More Accounts
OpenAI is rolling out its highly anticipated Advanced Voice feature to all ChatGPT Plus and Team users in select regions. The update will be rolled out throughout the week, so users in the regions where it's coming may need to wait for their turn. The standout feature in this update is the addition of five new voices. These aren't just regular sound English; the Advanced Voice comes with improved accents in select foreign languages, which should improve the clarity of interactions with users. This includes changes to the conversation speed and smoothness to make each conversation sound more natural. There is also a custom instructions feature that lets Plus and Team users tailor ChatGPT's behavior and responses, which sounds similar to Gems on Gemini, but it's a feature that affects all chats. It's like an area where users can tell ChatGPT to speak more slowly or quickly, how to refer to the user, and more. It's essentially a way to make it feel more personable and clear for each user. Finally, the Memory feature lets the AI recall previous conversations and maintain context over extended use. There's no need to wait by your ChatGPT and refresh, Plus and Team users will receive an in-app notification once they have access to Advanced Voice. A notable (and somewhat funny) feature is the ability for Advanced Voice to say the phrase "Sorry I'm late" in over 50 languages. This is to acknowledge the feature's delayed release and how much later it will be delayed for other users. Advanced Voice is not yet available in certain regions, including the EU, UK, Switzerland, Iceland, Norway, and Liechtenstein. This should feel like a new delay for people in those regions, but while some regions are currently excluded from this upgrade, we don't doubt that it is on the roadmap. OpenAI has not given a timeline for availability in these locations. Source: ChatGPT
[13]
OpenAI Begins Rollout of Advanced Voice to All Plus and Team Subscribers | PYMNTS.com
"While you've been patiently waiting, we've added Custom Instructions, Memory, five new voices, and improved accents," the company said in a Tuesday (Sept. 24) post on X. The feature is not yet available in the European Union, the United Kingdom, Switzerland, Iceland, Norway and Liechtenstein, OpenAI added in another post. Users can now choose from nine "lifelike output voices" for ChatGPT, with different tones and characters like "easygoing and versatile" and "animated and earnest," according to the company's Voice mode FAQ. It was reported July 30 that OpenAI was rolling out the alpha version of Advanced Voice Mode to a select group of ChatGPT Plus subscribers at that time and planned to begin a broader rollout to all premium users in the fall. To mitigate potential misuse of the feature, the company said at the time that it limited Advanced Voice Mode to preset voices created in collaboration with paid voice actors, so that it can't be used to impersonate specific individuals or public figures; implemented guardrails to block requests for violent or copyrighted content; and included filters to block requests for generating music or copyrighted audio, a move likely influenced by music industry legal actions against artificial intelligence (AI) companies. OpenAI had planned to roll the voice feature out in alpha in late June but said June 25 that it needed another month to do so. "For example, we're improving the model's ability to detect and refuse certain content," the company said at the time. "We're also working on improving the user experience and preparing our infrastructure to scale to millions while maintaining real-time responses." Many U.S. consumers are willing to pay for smart, reliable voice assistants, according to the PYMNTS Intelligence report, "How Consumers Want to Live in the Voice Economy." Twenty-two percent of Gen Z consumers are "highly willing" to pay more than $10 per month for a voice assistant that is as smart and reliable as a real person, per the report. The report also found that 54% of consumers would prefer using voice technology because it is faster than typing or using a touchscreen.
[14]
OpenAI's Advanced Voice mode is unavailable in the EU, and now we might know why
A tweet by OpenAI on X.com states simply, "Advanced Voice is not yet available in the EU, the UK, Switzerland, Iceland, Norway, and Liechtenstein." In response, a tweet by X.com user Dean W Ball highlights a section of the EU AI Act, a piece of the EU legislation that restricts "the placing on the market, the putting into service for this specific purpose, or the use of AI systems to infer emotions of a natural person." This would mean that ChatGPT Advanced Voice mode would be illegal in EU workplaces and schools because it can recognize emotions in the user's voice. We have to consider that OpenAI may be granted an exception of some kind, but as it stands, the AI Act will certainly keep European countries behind the rest of the world as AI advances. Of course, there is the argument of respecting the spirit of the law, rather than the letter of the law to consider. Still, it's certainly true that since Advanced Voice mode can sense and respond to your emotions, it would currently violate this condition. With Advanced Voice mode, you can talk to ChatGPT using your smartphone and get responses that make it feel like a real conversation with a human. To make it sound natural, Advanced Voice mode uses the GPT-4o LLM for faster responses and can interpret text, vision, and audio inputs. For example, you can ask it to look at something with your smartphone's camera and give an opinion. It can also remember things about you, making your interactions more of a sustained conversation over time. Additionally, there have been some improvements to Advanced Voice mode since we saw the first Alpha version. Now you get five new voices called Arbor, Maple, Sol, Spruce, and Vale, and there's a new blue orb visual effect as it talks. Accents in foreign languages have also been improved: Advanced Voice mode is only available to ChatGPT Plus and Teams subscribers. The rollout will be gradual but should be finished by the end of the week. You'll know when you've got it because you'll see a pop-up message next to the entry point to Voice Mode. But, as we mentioned previously, Advanced Voice is not yet available in the EU, the UK, Switzerland, Iceland, Norway, and Liechtenstein.
Share
Share
Copy Link
OpenAI has finally released its advanced voice feature for ChatGPT Plus and Team users, allowing for more natural conversations with the AI. The feature was initially paused due to concerns over potential misuse.
OpenAI has officially launched its advanced voice feature for ChatGPT Plus and Team users, marking a significant enhancement in AI-human interaction 1. This long-awaited feature enables users to engage in more natural, voice-based conversations with the AI, bridging the gap between text-based interactions and human-like dialogue.
The voice feature is now accessible to ChatGPT Plus subscribers and ChatGPT Team users 2. To utilize this capability, users need to activate it through the settings menu in their ChatGPT mobile app. Once enabled, a headphones icon appears, allowing users to initiate voice conversations with a simple tap.
OpenAI's voice feature employs advanced text-to-speech technology, offering users a choice of five distinct voices for their AI interactions 3. This diversity in voice options enhances the personalization of the user experience, making conversations feel more tailored and engaging.
The release of this feature comes after a months-long pause, initially announced in September but delayed due to concerns over potential misuse 4. OpenAI took this time to implement additional safety measures, addressing worries about the technology's potential for creating deepfakes or misleading audio content.
This advancement in voice interaction technology represents a significant step towards more intuitive and accessible AI interfaces. It opens up new possibilities for applications in various fields, including education, customer service, and accessibility for visually impaired users 5.
Initial user feedback has been largely positive, with many praising the natural flow of conversations and the convenience of voice interactions. However, some users have reported occasional glitches and limitations, indicating that while impressive, the technology is still evolving.
OpenAI has assured users that voice data will be handled with strict privacy measures. The company states that voice conversations are not recorded or stored, and the audio is processed in real-time to generate responses, maintaining user confidentiality.
Reference
[1]
[2]
[3]
OpenAI has rolled out an advanced voice mode for ChatGPT, allowing users to engage in verbal conversations with the AI. This feature is being gradually introduced to paid subscribers, starting with Plus and Enterprise users in the United States.
12 Sources
OpenAI launches a new voice-based interaction feature for ChatGPT Plus subscribers, allowing users to engage in conversations with the AI using voice commands and receive spoken responses.
29 Sources
OpenAI introduces an advanced voice mode for ChatGPT, allowing users to have spoken conversations with the AI. This feature is currently available for Plus and Enterprise users on iOS and Android devices.
2 Sources
OpenAI brings ChatGPT's Advanced Voice Mode to Windows and Mac desktop apps, offering users a more natural and intuitive way to interact with AI through voice conversations.
6 Sources
ChatGPT's new Advanced Voice Mode brings human-like speech to AI interactions, offering multilingual support, customization, and diverse applications across personal and professional domains.
2 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2024 TheOutpost.AI All rights reserved