8 Sources
8 Sources
[1]
The debut of Gemini 3.1 Flash Live could make it harder to know if you're talking to a robot
Text generated by artificial intelligence often has a particular vibe that gives it away as machine-generated, but it has become harder to pick out those idiosyncrasies as the tech has improved. We may be seeing a similar evolution of generative AI audio. Google has announced a new AI audio model called Gemini 3.1 Flash Live -- as the name implies, it's designed for real-time conversation. It's rolling out in some Google products starting today, and developers will be able to start building their own chatty robots with the model, too. Google says this AI is much faster and produces speech with a more natural cadence, aiming to solve a long-running issue with AI-generated speech. Like a chatbot, there's always a delay between input and output in generative audio systems. Longer delays and unnatural inflection make conversations feel sluggish and harder to follow. Researchers generally believe 300 milliseconds of latency is about the limit for optimal speech perception, but Google has not specified any particular delay for Gemini 3.1 Flash Live. It just vaguely has the speed you need. But benchmark numbers? Google has plenty of those, which it claims show that 3.1 Flash Live will be a more reliable way to have audio-to-audio AI conversations. For example, a big gain in the ComplexFuncBench Audio shows the new model is better at complex, multi-step tasks. Gemini 3.1 Flash Live also tops the charts in the Big Bench Audio test, which evaluates reasoning with a set of 1,000 audio questions. Meanwhile, a strong showing in Scale AI's Audio MultiChallenge means the new Gemini model is more able to cope with hesitation and interruptions in the audio input. Although it outpaces other real-time audio models, Gemini 3.1 Flash Live only manages 36.1 percent in this test. Audio models that are not designed to operate conversationally can reach scores over 50 percent in the MultiChallenge. The upshot is that Gemini 3.1 Flash Live should sound more like a person, to the point that Google felt it was time to integrate AI flags. The outputs from this model will have SynthID watermarks, which are not perceptible to human listeners. However, they can be detected if someone were to try to pass off Gemini AI speech as the real deal. Google has partnered with companies like Home Depot, Verizon, and others to test the model. They all have glowing reports in the blog post on how well 3.1 Flash Live can mimic human speech. So the next AI assistant you encounter on a phone call might sound much more realistic. Maybe you'll even think you're talking to a person, and SynthID can't help with that. Developers can now access the model in AI Studio, the Gemini API, and Gemini Enterprise for Customer Experience. The latter is essentially a toolkit for agentic shopping. Gemini 3.1 Flash Live will be seen most prominently in Gemini Live and Search Live (a feature of AI Mode). The new conversational AI is rolling out in those products starting today.
[2]
Search Live with Gemini's latest model tries to keep up with your rapid-fire questions
It's available now in Gemini Live and Search Live, the latter of which is now available worldwide. Google is rolling out another new Gemini model. Gemini 3.1 Flash Live is meant to enable quicker and more natural-sounding AI voices, among other, less immediately tangible benefits, and it's available now in a number of places across the Google ecosystem. In a blog post, Google's detailed the improvements that come with Gemini 3.1 Flash Live, which the company describes as its "highest-quality audio and voice model yet." Most people will experience the new model in Search Live and Gemini Live, though it also comes with a number of purported improvements for developers and enterprise customers. Google says that Gemini 3.1 Flash Live makes for "more helpful and natural responses" in the conversational-style Gemini Live and Search Live interfaces, and also responds more quickly than the previous model. Gemini 3.1 Flash Live is "inherently multilingual," a characteristic Google says made a global expansion of Search Live possible. Search Live is now available in multiple languages in more than 200 territories around the world. The update also benefits AI developers, Google says, thanks to better performance. Gemini 3.1 Flash Live scores higher on a number of benchmark tests, though those types of improvements aren't likely to be appreciable from a consumer standpoint. Finally, 3.1 Flash Live purportedly makes for a less miserable experience when interacting with an AI customer service agent. Google says the new model is more able to discern pitch and pace, which lets it tweak its approach when it calculates a customer is getting confused or annoyed, though it's presumably still not as effective on the phone as a well trained worker. Gemini 3.1 Flash Live is out now. You can experience it in Gemini Live or Search Live starting today.
[3]
Gemini 3.1 Flash Live: Making audio AI more natural and reliable
This content is generated by Google AI. Generative AI is experimental Today, we're advancing Gemini's real-time dialogue capabilities with Gemini 3.1 Flash Live, our highest-quality audio and voice model yet. It delivers the speed and natural rhythm needed for the next generation of voice-first AI, offering a more intuitive experience for developers, enterprises and everyday users. 3.1 Flash Live is available across Google products: We've improved 3.1 Flash Live's overall quality, making it more reliable for developers and enterprises to build voice-first agents that can complete complex tasks at scale. On ComplexFuncBench Audio, a benchmark that captures multi-step function calling with various constraints, it leads with a score of 90.8% compared to our previous model.
[4]
Gemini Live just doubled its memory, and longer conversations finally work
Outside of the office, Josh can be found digging into the latest video games, fantasy books, or tinkering with the newest features in Windows. Gemini Live has expanded and evolved quite a bit since Google first introduced it as a true Google Assistant replacement. From helping with daily problems to analyzing information on your screen, Gemini Live continues to be a solid tool to utilize, especially if you have an Android phone. And with the latest update to Gemini 3.1 Flash Live, Google says the assistant is about to get even better. Related Gemini isn't as useless as it was when you tried it two years ago AI that I first despised is now my Google Assistant replacement. Posts 3 By Keval Shukla Faster and more natural responses The classic upgrade claim One of the biggest upgrades that Google focused on in the announcement for Gemini 3.1 Flash Live is the fact that the AI model should be able to respond much faster and more precisely than it has in the past. This should make it easier to use Gemini Live to get instant help with your homework or just to ask it random questions. Additionally, Google claims that responses from Gemini Live should now be more natural overall, whether you're just asking it questions about your day or looking to dig into more complicated conversations or topics. The company claims that it's the "biggest upgrade yet" to the model, though that is often a claim we hear with many of these AI updates, and each update is, of course, going to bring more natural responses, since that's the goal. What really matters here is how useful the changes actually are when you start using them. Which is why one of the most notable things that Gemini Live fans will want to be aware of is the additional context parameters that Gemini 3.1 Flash Live brings to the table. Gemini 3.1 Flash Live can follow your conversation twice as long Talk longer before Gemini starts to spiral One of the biggest downfalls of AI models has always been contextual awareness. Because these models are essentially just computer systems, they can only follow a specific amount of data before the information begins to be overwritten. When that happens, conversations with the AI can degrade rapidly, as its responses start to lose much of the context that has helped it carry the conversation forward. These limitations can make it difficult to do some of our favorite things with Gemini Live. However, with the upgrade to Gemini 3.1 Flash Live, Google says that Gemini Live's context window has been increased two-fold, allowing it to hold onto a conversation thread twice as long. No exact numbers were provided here, but it will hopefully at least make it easier to hold longer brainstorming sessions with Gemini Live. This should also hopefully make it easier to take advantage of the different things you can do with the assistant, like hands-free meal planning and more, especially if you already own one of Google's Pixel devices, which has Gemini Live baked right into it. Google Pixel 10 Pro Brand Google SoC Tensor G5 Display 6.3" Super Actua display RAM 16 GB Storage 128 GB, 256 GB, 512 GB, 1 TB Battery 4870 or The Pixel 10 Pro is Google's highest-end flagship smartphone. It features an improved rear camera system, the Tensor G5 chip, seven years of software updates, and a 6.3" Super Actua display. $999 at Google Store $999 at Amazon $999 at Best Buy Expand Collapse Finally, Google is bringing all of these upgrades to everyone as part of Gemini Live and as its expansion for Search Live, which makes it easier to search the web using Gemini Live. The increased precision and contextual awareness, as well as its improved multilingual capacity should make it much easier for millions to use Search live in more than 200 different countries across the globe.
[5]
Google's Gemini Live Gets a Major Upgrade as Search Live Expands Globally
Google, on Thursday, announced two major updates to its artificial intelligence (AI)-powered live features. The Mountain View-based tech giant is now expanding Search Live globally, allowing users to receive relevant answers to their search queries using the Gemini assistant. Additionally, Gemini Live, the real-time conversational experience inside the Gemini app, is also getting an upgrade and will now be powered by the Gemini 3.1 Flash Live AI model. The newer model brings improvements such as faster and smarter responses and fewer awkward pauses. Google Focuses on Its AI-Powered Live Products In a post on X (formerly known as Twitter), the tech giant announced and detailed the upgrade to Gemini Live, which was earlier powered by the Gemini 2.5 AI model. Powered by the Gemini 3.1 Flash Live, it now gets faster responses and fewer awkward pauses. Additionally, Google also claims that the feature will provide users with smarter responses with twice the context window. It is also said to be able to adjust its answer lengths and tone to match the context. Google says the new AI model offers a more reliable choice for developers and enterprises to build voice-first agents that can complete complex tasks at scale. Gemini 3.1 Flash Live scores higher than Gemini 2.5 on the ComplexFuncBench Audio benchmark, which measures multi-step function calling, and the Audio MultiChallenge benchmark, which tests complex instruction following and long-horizon reasoning. All audio generated using the model is watermarked using SynthID. Notably, the same model also powers Search Live. Coming to Search Live, Google is now expanding the experience within the Google app in more than 200 countries and territories, or wherever AI Mode is available. It will also support all the languages Gemini currently does. Users can access the feature using both voice and camera. It can be activated by tapping the Live icon under the Search bar on both Android and iOS. Additionally, Search Live can also be accessed if the user is already pointing their camera with Google Lens by tapping the Live option at the bottom of the screen. This will open a real-time two-way conversation with the AI assistant, enabling users to interactively ask questions about their surroundings.
[6]
Google rolls out Gemini 3.1 Flash Live for real-time voice AI conversations, expands Search Live globally
Google has introduced Gemini 3.1 Flash Live, a real-time audio and voice AI model designed to enable faster, more natural conversational experiences. The model enhances latency, reliability, and dialogue quality for developers, enterprises, and everyday users, supporting the next generation of voice-first and multimodal AI applications. Gemini 3.1 Flash Live is built to handle real-time conversations with improved responsiveness and contextual understanding. It maintains natural dialogue flow while supporting multi-turn interactions, longer conversations, and dynamic user inputs. The model is designed to deliver reliable, natural-sounding conversation while completing complex tasks, with benchmarks demonstrating significant improvements over previous versions. For instance: Developers can use the Gemini Live API to build real-time conversational agents that process voice and visual inputs while responding instantly. Key capabilities include: Example usage through the Google GenAI SDK allows asynchronous connection to audio sessions and real-time interactions. Search Live has expanded globally, now supporting users in over 200 countries and territories with AI Mode enabled. Gemini 3.1 Flash Live powers real-time voice and camera interactions for Search, making queries more natural and interactive. Key features of Search Live include: This allows users to perform tasks that require dynamic interaction, such as troubleshooting, learning, or exploring objects in real life. Gemini 3.1 Flash Live supports scalable infrastructure and partner integrations for production use: Companies such as Verizon, LiveKit, and The Home Depot report positive results using the model in conversational workflows. All audio generated includes a SynthID watermark, embedded imperceptibly into the output. This allows detection of AI-generated content, supporting transparency and helping reduce misinformation.
[7]
Google launches Gemini 3.1 Flash Live audio model for developers By Investing.com
Investing.com - Google announced Thursday the release of Gemini 3.1 Flash Live, a new audio and voice model designed to enable real-time dialogue with improved precision and lower latency. The model is available to developers in preview through the Gemini Live API in Google AI Studio, to enterprises via Gemini Enterprise for Customer Experience, and to consumers through Search Live and Gemini Live. The model scored 90.8% on ComplexFuncBench Audio, a benchmark measuring multi-step function calling with constraints. On Scale AI's Audio MultiChallenge, which tests complex instruction following and long-horizon reasoning amid real-world audio interruptions, Gemini 3.1 Flash Live achieved a score of 36.1% with "thinking" enabled. Companies including Verizon (NYSE:VZ), LiveKit and The Home Depot (NYSE:HD) have provided positive feedback on the model's performance in their workflows. The model features improved tonal understanding to recognize acoustic nuances such as pitch and pace, and can dynamically adjust responses to users' expressions of frustration or confusion. In consumer applications, Gemini Live delivers faster responses than the previous model and can maintain conversation context for twice as long. The 3.1 Flash Live model supports the global expansion of Search Live, which is now available in more than 200 countries and territories with multilingual capabilities. All audio generated by 3.1 Flash Live includes SynthID watermarking, an imperceptible marker embedded in the audio output to enable detection of AI-generated content. Google stated the watermarking technology is designed to help prevent misinformation. This article was generated with the support of AI and reviewed by an editor. For more information see our T&C.
[8]
Google introduces Gemini 3.1 Flash Live AI model: Check features and availability
Gemini 3.1 Flash Live is available starting today through the Gemini API and Google AI Studio. Google has introduced a new AI model dubbed Gemini 3.1 Flash Live. According to the tech giant, the new model is built to help developers create AI agents that can see, hear, and respond to the world around them almost instantly. 'This is a step change in latency, reliability and more natural-sounding dialogue, delivering the quality needed for the next generation of voice-first AI,' Google said. The Gemini 3.1 Flash Live AI model is also said to better understand tone, emphasis and intent, enabling agents with key improvements. Gemini 3.1 Flash Live is designed to help developers build 'real-time voice and vision agents that can not only process the world around them, but also respond at the speed of conversation,' the tech giant said. One of the key improvements is better performance in noisy environments. The model can more accurately distinguish between a user's voice and background sounds such as traffic, television, or other environmental noise. This allows AI agents to be responsive and work even in real-world settings where there are distractions. Also read: Google releases Lyria 3 Pro AI model with longer music generation: How to access Google has also improved the instruction-following abilities. Gemini 3.1 Flash Live is better at following the instructions given by developers. Another key improvement is more natural and low-latency dialogue. As the tech giant explained, the new model 'improves on latency and is even more effective at recognising acoustic nuances like pitch and pace compared to 2.5 Flash Native Audio, making real-time conversations feel a lot more fluid and natural.' Also read: Meta to lay off hundreds of employees across departments amid AI push: Report Furthermore, Gemini 3.1 Flash Live supports more than 90 languages for real-time multi-modal conversations. Gemini 3.1 Flash Live is available starting today through the Gemini API and Google AI Studio. Developers can use the Gemini Live API to integrate the new model into their applications.
Share
Share
Copy Link
Google unveiled Gemini 3.1 Flash Live, its highest-quality audio and voice AI model designed for real-time conversations. The update brings faster responses, more natural cadence, and a doubled context window to Gemini Live and Search Live, which now expands to over 200 countries. All outputs include SynthID watermarks to identify AI-generated speech.
Google has announced Gemini 3.1 Flash Live, positioning it as the company's highest-quality AI audio model designed specifically for real-time conversations
1
3
. The new AI model delivers faster responses and more natural cadence, addressing long-standing issues with AI-generated speech that have made conversations feel sluggish and harder to follow1
. The update rolls out today across multiple Google products, including Gemini Live and Search Live, while developers gain access through AI Studio, the Gemini API, and Gemini Enterprise for Customer Experience1
.
Source: Ars Technica
While researchers generally believe 300 milliseconds of latency is optimal for speech perception, Google has not specified exact delay numbers for the new model, stating only that it has "the speed you need"
1
. The company emphasizes that Gemini 3.1 Flash Live makes for "more helpful and natural responses" in conversational-style interfaces2
.Google has backed its claims with substantial benchmark scores demonstrating improved reliability for voice-first AI experiences. On ComplexFuncBench Audio, which measures multi-step function calling with various constraints, Gemini 3.1 Flash Live achieves a score of 90.8 percent compared to previous models
3
. The AI model also tops charts in the Big Bench Audio test, which evaluates reasoning with a set of 1,000 audio questions1
.In Scale AI's Audio MultiChallenge, which tests the ability to handle conversational interruptions and hesitation, Gemini 3.1 Flash Live scores 36.1 percent
1
. While this outpaces other real-time audio models, non-conversational audio models can reach scores over 50 percent in the same test, suggesting room for improvement in handling natural speech patterns.One of the most significant upgrades comes in the form of an expanded context window, which has been increased two-fold
4
. This enhancement addresses a critical limitation in conversational AI, where models can only follow a specific amount of data before information begins to be overwritten. When that happens, conversations degrade rapidly as responses lose the context that helps carry the dialogue forward4
.Source: Android Authority
The doubled context window allows Gemini Live to hold onto conversation threads twice as long, making it easier to conduct extended brainstorming sessions and complex multi-turn dialogues
4
. Google claims the feature can now adjust answer lengths and tone to match context more effectively5
. The AI model is also "inherently multilingual," a characteristic that enabled the global expansion of Search Live.As natural-sounding AI voices become increasingly difficult to distinguish from human speech, Google has integrated SynthID watermarks into all audio generated by Gemini 3.1 Flash Live
1
5
. These watermarks are not perceptible to human listeners but can be detected if someone attempts to pass off AI-generated speech as authentic human voice1
.However, this protection has limitations. While SynthID can identify AI-generated audio after the fact, it cannot help users determine in real-time whether they're speaking with an AI assistant or a human during a phone call
1
. This raises questions about transparency in AI-powered customer service interactions.Related Stories
Google has partnered with enterprise clients including Home Depot and Verizon to test the model, with all reporting positive experiences in how well Gemini 3.1 Flash Live can mimic human speech
1
. For customer service agents, the new AI model can better discern pitch and pace, allowing it to adjust its approach when it calculates a customer is getting confused or annoyed2
.Developers can now access Gemini 3.1 Flash Live to build voice-first agents capable of completing complex tasks at scale
3
. The model is available through AI Studio, the Gemini API, and Gemini Enterprise for Customer Experience, which serves as a toolkit for agentic shopping applications1
.Alongside the Gemini 3.1 Flash Live announcement, Google is expanding Search Live globally to more than 200 countries and territories wherever AI Mode is available
5
. The feature supports all languages currently available in Gemini and can be accessed via voice and camera on both Android and iOS devices5
.
Source: Gadgets 360
Users can activate Search Live by tapping the Live icon under the search bar in the Google app, or by tapping the Live option while using Google Lens to ask questions about their surroundings in real-time
5
. This expansion makes AI-powered live features accessible to millions of users worldwide, potentially transforming how people interact with search technology and AI assistants in their daily lives.Summarized by
Navi
[1]
[2]
13 Nov 2025•Technology

12 Dec 2025•Technology

21 Aug 2025•Technology

1
Technology

2
Policy and Regulation

3
Policy and Regulation
