37 Sources
37 Sources
[1]
Google releases Gemini 3 Flash, promising improved intelligence and efficiency
Google began its transition to Gemini 3 a few weeks ago with the launch of the Pro model, and the arrival of Gemini 3 Flash kicks it into high gear. The new, faster Gemini 3 model is coming to the Gemini app and search, and developers will be able to access it immediately via the Gemini API, Vertex AI, AI Studio, and Antigravity. Google's bigger gen AI model is also picking up steam, with both Gemini 3 Pro and its image component (Nano Banana Pro) expanding in search. This may come as a shock, but Google says Gemini 3 Flash is faster and more capable than its previous base model. As usual, Google has a raft of benchmark numbers that show modest improvements for the new model. It bests the old 2.5 Flash in basic academic and reasoning tests like GPQA Diamond and MMMU Pro (where it even beats 3 Pro). It gets a larger boost in Humanity's Last Exam (HLE), which tests advanced domain-specific knowledge. Gemini 3 Flash has tripled the old models' score in HLE, landing at 33.7 percent without tool use. That's just a few points behind the Gemini 3 Pro model. Google is talking up Gemini 3 Flash's coding skills, and the provided benchmarks seem to back that talk up. Over the past year, Google has mostly pushed its Pro models as the best for generating code, but 3 Flash has done a lot of catching up. In the popular SWE-Bench Verified test, Gemini 3 Flash has gained almost 20 points on the 2.5 branch. The new model is also a lot less likely to get general-knowledge questions wrong. In the Simple QA Verified test, Gemini 3 Flash scored 68.7 percent, which is only a little below Gemini 3 Pro. The last Flash model scored just 28.1 percent on that test. At least as far as the evaluation scores go, Gemini 3 Flash performs much closer to Google's Pro model versus the older 2.5 family. At the same time, it's considerably more efficient, according to Google. One of Gemini 3 Pro's defining advances was its ability to generate interactive simulations and multimodal content. Gemini 3 Flash reportedly retains that underlying capability. Gemini 3 Flash offers better performance than Gemini 2.5 Pro did, but it runs workloads three times faster. It's also a lot cheaper than the Pro models if you're paying per token. One million input tokens for 3 Flash will run devs $0.50, and a million output tokens will cost $3. However, that's an increase compared to Gemini 2.5 Flash input and output at $0.30 and $2.50, respectively. The Pro model's tokens are $2 (1M input) and $12 (1M output). Simplified model selection Google's rapid-fire release of new AI models and tools has occasionally made the Gemini app a bit confusing. Over recent weeks, the settings have been pared down and rearranged. With the release of Gemini 3 Flash, that will become the new default model in the Gemini app and web interface -- that's the Fast setting in the app, as well as the one labeled Thinking, which uses simulated reasoning for better outputs. Gemini 3 Pro will continue to be available under the Pro option. That's still a bit misleading, though, as both versions of Gemini 3 can use the reasoning process that Google likes to call "thinking" to generate answers. Whichever one you choose in the app, you can then select tools like image generation, canvas, and Deep Research. In addition to its debut in the Gemini app, the new Flash model will be coming to search immediately. When Google says "search" in this context, it mostly means AI Mode. Gemini 3 Flash will be the default model in AI Mode going forward. That means free users will see a notable improvement when using the Gemini app. There are no specific changes to AI Overviews. Google says AI Overviews will continue to use the best model for the job. Due to its place at the top of organic search results, though, you'll probably see it lean on less capable (but faster) models. Gemini 3 Flash could show up there -- even Gemini 3 Pro could power some complex queries in AI Overviews for paying subscribers. Gemini 3 Pro is also expanding in AI Mode for all US-based users. Likewise, Gemini 3 Pro Image (Nano Banana Pro) will also arrive in AI mode for all. There will be limits on free access to these models, but Google hasn't specified what those are. It does say that Pro and Ultra subscribers will enjoy much higher usage limits.
[2]
Google launches Gemini 3 Flash, makes it the default model in the Gemini app | TechCrunch
Google today released its fast and cheap Gemini 3 Flash model, based on the Gemini 3 released last month, looking to steal OpenAI's thunder. The company is also making this the default model in the Gemini app and AI mode in search. The new Flash model arrives six months after Google announced the Gemini 2.5 Flash model, offering significant improvements. On the benchmark, the Gemini 3 Flash model outperforms its predecessor by a significant margin and matches the performance of other frontier models, like Gemini 3 Pro and GPT 5.2, in some measures. For instance, it scored 33.7% without tool use on Humanity's Last Exam benchmark, which is designed to test expertise across different domains. In comparison, Gemini 3 Pro scored 37.5%, Gemini 2.5 Flash scored 11%, and the newly released GPT-5.2 scored 34.5%. On the multimodality and reasoning benchmark MMMU-Pro, the new model outscored all competitors with an 81.2% score. Google is making Gemini 3 Flash the default model in the Gemini app globally, replacing Gemini 2.5 Flash. Users can still choose the Pro model from the model picker for math and coding questions. The company says the new model is good at identifying multimodal content and giving you an answer based on that. For instance, you can upload your pickleball short video and ask for tips; you can try drawing a sketch and have the model guess what you are drawing; or you can upload an audio recording to get analysis or generate a quiz. The company also said the model better understands the intent of users' queries and can generate more visual answers with elements like images and tables. You can also use the new model to create app prototypes in the Gemini app using prompts. The Gemini 3 Pro is now available to everyone in the U.S for search and more people in the U.S. can access the Nano Banana Pro image model in search, as well. Google noted that companies like JetBrains, Figma, Cursor, Harvey, and Latitude are already using the Gemini 3 Flash model, which is available through Vertex AI and Gemini Enterprise. For developers, the company is making the model available in a preview model through the API and in Antigravity, Google's new coding tool released last month. The company said the Gemini 3 Pro scores 78% on the SWE-bench verified coding benchmark, only outperformed by GPT-5.2. It added that the model is ideal for video analysis, data extraction, and visual Q&A, and because of its speed, it is suited for quick and repeatable workflows. Model pricing is $0.50 per 1 million input tokens and $3.00 per 1 million output tokens. This is slightly more expensive than $0.30 per 1 million input tokens and $2.50 per 1 million output tokens of Gemini Flash 2.5. But Google claims that the new model outperforms the Gemini 2.5 Pro model while being three times faster. And, for thinking tasks, it uses 30% fewer tokens on average than 2.5 Pro. That means overall, you might save on the number of tokens for certain tasks. "We really position flash as more of your workhorse model. So if you look at, for example, even the input and output prices at the top of this table, Flash is just a much cheaper offering from an input and output price perspective. And so it actually allows for, for many companies, bulk tasks," Tulsee Doshi, Senior Director & Head of Product for Gemini Models, told TechCrunch in a briefing Since it released Gemini 3, Google has processed over 1 trillion tokens per day on its API, amid its fierce release and performance war with OpenAI. Earlier this month, Sam Altman reportedly sent an internal "Code Red" memo to the OpenAI team after ChatGPT's traffic dipped as Google's market share in consumers rose. Post that, OpenAI has released GPT-5.2 and a new image generation model. OpenAI also boasted about its growing enterprise use and said the ChatGPT messages volume has grown 8x since November 2024. While Google didn't directly address the competition with OpenAI, it said that the release of new models is challenging all companies to be active. "Just about what's happening across the industry is like all of these models are continuing to be awesome, challenge each other, push the frontier. And I think what's also awesome is as companies are releasing these models," Doshi said. "We're also introducing new benchmarks and new ways of evaluating these models. And so that's also encouraging us."
[3]
Google Says Its New Gemini 3 Flash AI Model Is Better and Faster Than 2.5 Pro
Imad is a senior reporter covering Google and internet culture. Hailing from Texas, Imad started his journalism career in 2013 and has amassed bylines with The New York Times, The Washington Post, ESPN, Tom's Guide and Wired, among others. Google's new Gemini 3 Flash AI model is built for faster output at a lower cost, and functions as well as previous powerful reasoning models, Google said in a press release on Wednesday. According to benchmark tests, Gemini 3 Flash achieves PhD-level reasoning in the GPQA Diamond test, with a score of 90.4%, and achieved a score of 33.7% (without tools) in Humanity's Last Exam. These are very difficult tests designed to push AI models and require expert-level knowledge. By comparison, Gemini 3 Pro achieved scores of 91.9% and 37.5%, respectively. Google said Gemini 3 Flash outperforms Gemini 2.5 Pro, which was Google's top model when it released earlier this year, at three times the speed. Don't miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source. Gemini 3 Flash is available in Google AI Studio and Gemini CLI for developers. For general consumers, it's rolling out in the Gemini app and AI Mode in Google Search. For enterprise users, Gemini 3 Flash is available in Vertex AI and Gemini Enterprise. Google said Gemini 3 Flash can be a handy agent for customer support or in-game support, tasks that require fast responses. "Gemini 3 Flash excels at grasping the nuance of your question to serve thoughtful, comprehensive responses that are visually digestible -- pulling real-time local information and helpful links from across the web," Google said in a blog post. Gemini 3 Pro and nano banana, Google's image model, are also landing in AI Mode in Search. They'll be accessible via a dropdown menu and is only available to AI Pro and Ultra subscribers.
[4]
You can try Google's new Gemini 3 Flash AI model today for free - it's even in Search's AI Mode
Google unveiled Gemini 3 Flash on Wednesday.It's the default model in AI Mode in Search and the Gemini app.It outperformed Gemini 3 Pro on agentic coding. Google has unveiled Gemini 3 Flash, the company's latest AI model, which it says balances speed with industry-leading reasoning and multimodal capabilities. Along with its global release, Gemini 3 Flash has also been made the default model in both Google Search, the Gemini app, and AI Mode. That's significant, since in the earlier days of the AI race, companies -- even the more powerful ones like Google -- would tend to roll out new AI tools slowly and cautiously, getting a sense of user feedback before adding them to their most popular and widely used services. Also: Inside the making of Gemini 3 - how Google's slow and steady approach won the AI race (for now) Last month, Google embedded the recently arrived Gemini 3 into Search; then, as now, the decision to move quickly and upgrade the company's cornerstone products with a brand new model points strongly to a strong sense of confidence within the company as it steps up its competition with OpenAI and other developers. "For too long, AI forced a choice: big models that were slow and expensive, or high-speed models that were less capable," Josh Woodward, VP of Google Labs and Gemini, wrote in a blog post published Wednesday. "Gemini 3 Flash ends this compromise. Gemini 3 Flash delivers smarts and speed." At this point in the AI race, you might be feeling a little motion sickness from the sheer number of new models and tools that have been dropping pretty much on a daily basis. Companies have an incentive to move quickly and promote every new AI product as a complete game-changer, since it adds to their aura of power. It's therefore always important to ask: What does this or that tool actually offer? Is it just the same old AI capabilities in a new guise, or is there something that's actually, substantively new, here? If you write code for a living (or just for fun), you're likely to see real benefits from Gemini 3 Flash. The model combines Gemini 3 Pro's frontier coding skill with low latency, and it also offers state-of-the-art agentic capabilities, "which means it can enable more intelligent applications -- like live customer support agents or in-game assistants -- that demand both quick answers and deep reasoning," Tulsee Doshi, a senior director of product management at Google, wrote in a blog post. Also: Google just rolled out Gemini 3 to Search - here's what it can do and how to try it For the rest of us, the global deployment of Gemini 3 Flash will primarily mean a boost in the perks they've already come to expect from Gemini app or AI Mode in Search. For example, non-coding experts can feed a stream of consciousness description about an app they'd like to build into the app, and Gemini will quickly spit out a prototype. Or someone running a search can ask for travel advice and reservations for a last-minute vacation, and the model will pull up results based on that user's particular preferences. Users will also probably notice a significant boost in the quality and flexibility of the multimodal capabilities offered by Gemini through AI Mode and the app: it can respond to questions about images, video, audio, or text, and thanks to Nano Banana Pro, generate images in AI Mode (just select "Thinking with 3 Pro" and then "Create Images Pro" in the model drop-down menu). The model fell just short of matching GPT-5.2's performance on Humanity's Last Exam, and it outperformed both Gemini 3 Pro and Claude Sonnet 4.5 on the agentic coding SWE-Bench Verified benchmark, according to data published by Google. Gemini 3 Flash is available now in preview through the Gemini API in Google AI Studio, and it's starting to roll out on Wednesday for free to all users in the Gemini app and in AI Mode in Search.
[5]
Gemini 3 Flash is here, bringing a 'huge' upgrade to the Gemini app
Google is upgrading its Gemini app with Gemini 3 Flash, a more efficient version of its flagship AI model that's still capable of fielding complex requests. The new model will replace Gemini 2.5 Flash as the default inside the Gemini app, and it's headed to Google Search, too. Its arrival comes one month after the launch of Gemini 3 Pro, which showed advancements in reasoning, coding, and its ability to process images, text, and videos simultaneously. Google says Gemini 3 Flash "retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost." Tulsee Doshi, Google DeepMind's senior director and head of product, tells The Verge that the jump to Gemini 3 Flash will be a "huge upgrade" for most users. "With Gemini 3 Flash... it'll be a faster turnaround from a latency perspective," Doshi says, adding that you'll also see "more detailed, nuanced answers" when compared to Gemini 2.5 Flash. Gemini 3 Flash also outperforms the last-gen flagship, Gemini 2.5 Pro, while operating at a "fraction of the cost," according to Google. As an example, the company says Gemini 3 Flash can generate a plan based on a series of videos and images in "just a few seconds." In addition to launching inside the Gemini app globally, Gemini 3 Flash is becoming the default model powering AI Mode in Google Search, which previously ran on 2.5 Flash. Google is bringing Gemini 3 Flash to developers as well. It's coming to Google AI Studio, the Gemini API, Google Antigravity, Gemini CLI, Android Studio, and Vertex AI.
[6]
Google's Gemini 3 Flash model outperforms GPT-5.2 in some benchmarks
Almost exactly a month after the debut of Gemini 3 Pro in November, Google has begun rolling out the more efficient Flash version of its latest AI model. According to the company, the new system offers similar "pro-grade" reasoning performance as its flagship model at a fraction of the cost, making it ideal for everyday use. In benchmarks, the new system performed significantly better than Google's previous generation models, including Gemini 2.5 Pro. More notably, in Google's testing it managed to trade blows with GPT-5.2, the model OpenAI rushed out to counter Gemini 3 Pro. For example, in the tough Humanity's Last Exam suite, Gemini 3 Flash scored less than a percentage point worse than GPT-5.2 when neither model had access to tools like web search. In a handful of other benchmarks, Google's more efficient system even managed to outperform OpenAI's latest. For instance, in MMMU-Pro, a benchmark designed to test a model's multimodal understanding and reasoning, it edged out GPT-5.2 with a result of 81.2 percent compared to 79.5 percent. Of course, benchmarks only tell a partial story; we'll have to see what people think once both systems are broadly available. Still, the fact Gemini 3 Flash is even close to GPT-5.2, and the "Extra High" reasoning mode at that, is a worrying sign for OpenAI. As with Gemini 3 Pro, Google is rolling out the new model to both the Gemini App and AI Mode in Search where it will be the default model for both services. "That means all of our Gemini users globally will get access to the Gemini 3 experience at no cost, giving their everyday tasks a major upgrade," Google explains. While on the subject of AI Mode, it's now possible to access Nano Banana Pro, Google's latest image generator, directly from the chatbot. Provided you live in the US, select "Thinking with 3 Pro," followed by "Create Images Pro" from the model picker.
[7]
Google AI Mode gets a speed boost with Gemini 3 Flash
The new model is more performant than 2.5 Pro while working faster than 2.5 Flash. The Gemini app and Google's AI Mode should feel a little quicker starting today. Google's announced that its latest Gemini model, Gemini 3 Flash, is now rolling out across its various AI products. As the name implies, 3 Flash is meant to be quicker than other Gemini 3 models while still offering strong performance on AI benchmarks. Google announced the rollout in a blog post this morning. The company says that Gemini 3 Flash includes the best of Gemini 3's "Pro-grade reasoning" while operating with less latency and, crucially, at a lower cost, an area that's still a sore spot for AI providers at large. The main takeaway for average users is that both the Gemini app and AI Mode in Search should be able to provide answers more quickly, particularly when queries are complex. Gemini 3 Flash handily outperforms both the previous 2.5 Flash and even 2.5 Pro on a wide variety of benchmarks while working three times as fast as the older Pro model. Gemini 3 Flash is apparently not as inexpensive to operate as 2.5 Flash; Google's data shows 3 Flash's input price is rated at 50 cents per one million tokens, up from 30 cents in the previous version. Google does go on to say that Gemini 3 Flash "uses 30% fewer tokens on average than 2.5 Pro," so that increase in per-token cost could be offset to some extent. Google says that Gemini 3 Flash offers enticing improvements for developers, including "better coding and agent capabilities" that apparently outperform Gemini 3 Pro. Gemini 3 Flash is available now in Google AI Studio, Google Antigravity, and Android Studio. Gemini 3 Flash is now the default model for Search's AI Mode for all users. Google says this allows AI Mode to answer "more nuanced questions" with "thoughtful" responses. If you need more out of AI Mode -- custom interactive tools or specialized visuals, for example -- you're able to select Gemini 3 Pro from the model drop-down. With Gemini 3 Pro selected, you can also use Nano Banana Pro to create images and infographics. You can try out Gemini 3 Flash in AI Mode and the Gemini app right now.
[8]
Google just upgraded your Gemini app with a faster default model
Exactly one month after Google rolled out the heavyweight Gemini 3 Pro flagship AI model, it is now ready to bring Gemini 3's more efficient 'Flash' variant to the masses. Replacing Gemini 2.5 Flash starting today, Google's latest offering is now starting to show up for users in the Gemini app and AI Mode in Search. Related Google's latest AI model invades Search on day one Gemini 3 is here Posts By Karandeep Singh Oberoi Designed to bridge the gap between "frontier-level" capabilities and a fast response time, Gemini 3 Flash, despite being built for speed, punches well above its weight class. "With Gemini 3, we introduced frontier performance across complex reasoning, multimodal and vision understanding and agentic and vibe coding tasks. Gemini 3 Flash retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost," wrote the tech giant. The model outperforms the likes of Gemini 3 Pro, Gemini 2.5 Flash, Gemini 2.5 Pro, Claude Sonnet 4.5, GPT-5.2, and Grok 4.1 Fast when it comes to multimodal understanding and reasoning. Similarly, it also outperformed or tied with all the above-mentioned models when it came to multilingual Q&A and long horizon real-world software tasks. Elsewhere, the model can clearly hold its own with several other benchmarks, including Humanity's Last Exam, scientific knowledge, math, agentic coding, and more. The efficient model outperforms Gemini 2.5 Pro in all benchmark tests, all while using 30 percent fewer tokens at $0.50 per 1M input tokens. Gemini 3 Flash is now the default model in the Gemini app. All of our Gemini users globally will get access to the Gemini 3 experience at no cost, giving their everyday tasks a major upgrade. Gemini 3 Flash unlocks a faster Search experience Gemini 3 Flash can do everything that 2.5 Flash did, but better, complete with Gemini 3's Pro-grade reasoning. This allows the model to analyze short video content and help you come up with plans, analyze audio recordings and identify knowledge gaps, create a custom quiz, give you detailed explanations, and more. In addition to the Gemini app, Gemini 3 Flash is also starting to roll out as the default model for AI Mode in Search. Google says that the new model should allow AI Mode to better understand nuances in your questions. "It considers each aspect of your query to serve thoughtful, comprehensive responses that are visually digestible -- pulling real-time local information and helpful links from across the web," wrote the tech giant. Apart from AI Mode and the Gemini app, Gemini 3 Flash is also available to developers and professionals via the Gemini API in Google AI Studio, Google Antigravity, Vertex AI, Gemini Enterprise, Gemini CLI, and Android Studio.
[9]
Google launches Gemini 3 Flash as faster, cheaper default AI model
The battle for AI dominance is playing out in real time. A day after OpenAI launched GPT Image 1.5, Google rolled out Gemini 3 Flash, pushing it as the default model across its consumer apps and Search. The release marks Google's latest attempt to challenge OpenAI, as the AI race intensifies across consumer apps, search, and enterprise software. The company is making Gemini 3 Flash the default model in the Gemini app worldwide and in AI Mode within Google Search. It replaces Gemini 2.5 Flash, which launched just six months ago.
[10]
Google's new default AI model: Gemini 3 Flash is faster and stronger
This upgrade enhances Google's AI products with improved visual understanding, making advanced AI capabilities more accessible for everyday workflows and data analysis. Google has now launched Gemini 3 Flash, a faster and more cost-effective AI model based on Gemini 3. According to Google, Gemini 3 Flash is up to three times faster than Gemini 2.5 Flash, and it outperforms previous Flash models in all internal tests. In several benchmark tests, Gemini 3 Flash performed on par with both Gemini 3 Pro and OpenAI's GPT-5.2. In the multimodal test MMMU-Pro, it even topped the list with a result of 81.2 percent. The Flash model is supposed to be adapted for fast and repetitive workflows. According to Google, it's particularly suitable for video analysis, data extraction, and visual queries. It should also be better at understanding the user's intentions and can deliver more visual answers, for example with images and tables. Gemini 3 Flash will now be the default model in both the Gemini app and Google's AI-powered search mode.
[11]
Gemini app adds 3 Flash 'Fast' and 'Thinking' models, with 3 'Pro'
With today's launch, Gemini 3 Flash is widely rolling out to the Gemini app and AI Mode in Search. This is one of Google's speediest model rollouts. Since the launch of 3 Pro in November, Gemini's model picker has offered: Fast - "Answers quickly" and Thinking with 3 Pro - "Thinks through complex topics." With today's launch, the model picker now has a "Gemini 3" label at the top: It's widely rolling out on gemini.google.com, and seeing pretty wide availability in the Gemini app as of this morning. Meanwhile, in the past day, you might have seen a "Now Gemini checks your screen when needed" message upon launching the overlay. This contextual awareness might save you from using "Ask about screen." When you say or type something like "Explain this," Gemini checks your screen and app info to give you more relevant answers. Turn this off in Settings anytime. Gemini 3 Flash is also globally rolling out to AI Mode as the default model. It's appearing as "Auto" in the model picker, which is also now available in the 'plus' menu. In the US, Google is bringing Gemini 3 Pro and Nano Banana Pro (Gemini 3 Pro Image) to all users. Select "Thinking with 3 Pro" from the dropdown. Google AI Pro and Ultra subscribers will get higher usage limits.
[12]
I test AI tools for a living and Gemini 3 Flash is my favorite free model of the year -- here's why
2025 has been a packed year for Google's Gemini. Every couple of months, the team has released new updates, features and some truly revolutionary ideas. But it was a recent, and seemingly smaller, update that has impressed me the most. Recently, Google announced the launch of Gemini Flash. This is an addition to the Gemini 3 family, offering a faster, more affordable alternative to the ultra-intelligent Gemini 3 Pro. Normally, these more affordable versions of AI models offer a clear trade-off. You get a faster response, which costs less to run, but it just can't perform as well as the more powerful alternatives. With Gemini 3 Flash, Google found a solution to this problem. It just made an AI model that is fast, as well as being incredibly smart. It can take on challenging tasks, but doesn't force you to spend ages waiting for a response. I've seen a lot of solid AI updates, but here's why it has a place as one of the best. Open to everyone Upon its release, Google made Gemini 3 Flash free for absolutely anyone to use. In fact, the company made this the default version of the tool. Compared to other companies, like OpenAI and Anthropic, this is great news. It means that, even for those who don't want to pay for AI, there is a version that can take on the vast majority of challenges, and won't put limits on your usage. Sure, for those who want more power out of their AI, there are two additional options. Google offers Gemini 3 Pro and a Deep Think mode for those willing to pay for one of its premium plans. But for everyone else, the cost barrier to good, fast AI has been removed with this update. A more than capable choice Despite being both free, and positioned as the speedy workhorse of the Gemini family, Flash is undeniably impressive. Since it came out, I've been using Gemini Flash frequently. It produces really fast responses, without leaving out details. In fact, I have been consistently blown away with just how much it can provide in a limited amount of time. A couple of days ago, I had Gemini 3 Flash read through a 75-page document, and offer up its recommendations for next steps. It returned a well-thought through response, along with costs, plans for the future and concerns from the document. However, it's not just my own experience of it. On launch, Google posted a list of benchmarking scores for the model. This showed Gemini 3 Flash scoring at similar levels to its competitors, falling only slightly behind the likes of GPT-5.2, Claude Sonnet 4.5 and Grok 4.1 in a variety of challenging tests. Multimodal capabilities With the release of Gemini 3, Google made the point of offering a vast improvement in the models ability to understand documents, spatial reasoning, screen records and videos. Surprisingly, Gemini 3 Flash continues this same concept, able to understand inputs while still working at faster speeds. This is why it can handle large documents, or watch long videos and report back on its findings surprisingly easily. Not only that, but Gemini 3 Flash can handle multiple sources at once. If you're asking for tips on a garden redesign, you can include measurements, videos and images, as well as descriptions to get the best responses. Compared to other fast models, like GPT-5.2 Fast or Grok 4.1 Fast, this is a great additional feature, rarely seen at this level. For most tasks, Gemini Flash will prove to be more than capable for the average person's needs, often going above and beyond. Follow Tom's Guide on Google News and add us as a preferred source to get our up-to-date news, analysis, and reviews in your feeds.
[13]
Gemini 3 Flash arrives with reduced costs and latency -- a powerful combo for enterprises
Enterprises can now harness the power of a large language model that's near that of the state-of-the-art Google's Gemini 3 Pro, but at a fraction of the cost and with increased speed, thanks to the newly released Gemini 3 Flash. The model joins the flagship Gemini 3 Pro, Gemini 3 Deep Think, and Gemini Agent, all of which were announced and released last month. Gemini 3 Flash, now available on Gemini Enterprise, Google Antigravity, Gemini CLI, AI Studio, and on preview in Vertex AI, processes information in near real-time and helps build quick, responsive agentic applications. The company said in a blog post that Gemini 3 Flash "builds on the model series that developers and enterprises already love, optimized for high-frequency workflows that demand speed, without sacrificing quality. The model is also the default for AI Mode on Google Search and the Gemini application. Tulsee Doshi, senior director, product management on the Gemini team, said in a separate blog post that the model "demonstrates that speed and scale don't have to come at the cost of intelligence." "Gemini 3 Flash is made for iterative development, offering Gemini 3's Pro-grade coding performance with low latency -- it's able to reason and solve tasks quickly in high-frequency workflows," Doshi said. "It strikes an ideal balance for agentic coding, production-ready systems and responsive interactive applications." Early adoption by specialized firms proves the model's reliability in high-stakes fields. Harvey, an AI platform for law firms, reported a 7% jump in reasoning on their internal 'BigLaw Bench,' while Resemble AI discovered that Gemini 3 Flash could process complex forensic data for deepfake detection 4x faster than Gemini 2.5 Pro. These aren't just speed gains; they are enabling 'near real-time' workflows that were previously impossible. More efficient at a lower cost Enterprise AI builders have become more aware of the cost of running AI models, especially as they try to convince stakeholders to put more budget into agentic workflows that run on expensive models. Organizations have turned to smaller or distilled models, focusing on open models or other research and prompting techniques to help manage bloated AI costs. For enterprises, the biggest value proposition for Gemini 3 Flash is that it offers the same level of advanced multimodal capabilities, such as complex video analysis and data extraction, as its larger Gemini counterparts, but is far faster and cheaper. While Google's internal materials highlight a 3x speed increase over the 2.5 Pro series, data from independent benchmarking firm Artificial Analysis adds a layer of crucial nuance. In the latter organization's pre-release testing, Gemini 3 Flash Preview recorded a raw throughput of 218 output tokens per second. This makes it 22% slower than the previous 'non-reasoning' Gemini 2.5 Flash, but it is still significantly faster than rivals including OpenAI's GPT-5.1 high (125 t/s) and DeepSeek V3.2 reasoning (30 t/s). Most notably, Artificial Analysis crowned Gemini 3 Flash as the new leader in their AA-Omniscience knowledge benchmark, where it achieved the highest knowledge accuracy of any model tested to date. However, this intelligence comes with a 'reasoning tax': the model more than doubles its token usage compared to the 2.5 Flash series when tackling complex indexes. This high token density is offset by Google's aggressive pricing: when accessing through the Gemini API, Gemini 3 Flash costs $0.50 per 1 million input tokens, compared to $1.25/1M input tokens for Gemini 2.5 Pro, and $3/1M output tokens, compared to $ 10/1 M output tokens for Gemini 2.5 Pro. This allows Gemini 3 Flash to claim the title of the most cost-efficient model for its intelligence tier, despite being one of the most 'talkative' models in terms of raw token volume. Here's how it stacks up to rival LLM offerings: More ways to save But enterprise developers and users can cut costs further by eliminating the lag most larger models often have, which racks up token usage. Google said the model "is able to modulate how much it thinks," so that it uses more thinking and therefore more tokens for more complex tasks than for quick prompts. The company noted Gemini 3 Flash uses 30% fewer tokens than Gemini 2.5 Pro. To balance this new reasoning power with strict corporate latency requirements, Google has introduced a 'Thinking Level' parameter. Developers can toggle between 'Low' -- to minimize cost and latency for simple chat tasks -- and 'High' -- to maximize reasoning depth for complex data extraction. This granular control allows teams to build 'variable-speed' applications that only consume expensive 'thinking tokens' when a problem actually demands PhD-level lo The economic story extends beyond simple token prices. With the standard inclusion of Context Caching, enterprises processing massive, static datasets -- such as entire legal libraries or codebase repositories -- can see a 90% reduction in costs for repeated queries. When combined with the Batch API's 50% discount, the total cost of ownership for a Gemini-powered agent drops significantly below the threshold of competing frontier models "Gemini 3 Flash delivers exceptional performance on coding and agentic tasks combined with a lower price point, allowing teams to deploy sophisticated reasoning costs across high-volume processes without hitting barriers," Google said. By offering a model that delivers strong multimodal performance at a more affordable price, Google is making the case that enterprises concerned with controlling their AI spend should choose its models, especially Gemini 3 Flash. Strong benchmark performance But how does Gemini 3 Flash stack up against other models in terms of its performance? Doshi said the model achieved a score of 78% on the SWE-Bench Verified benchmark testing for coding agents, outperforming both the preceding Gemini 2.5 family and the newer Gemini 3 Pro itself! For enterprises, this means high-volume software maintenance and bug-fixing tasks can now be offloaded to a model that is both faster and cheaper than previous flagship models, without a degradation in code quality. The model also performed strongly on other benchmarks, scoring 81.2% on the MMMU Pro benchmark, comparable to Gemini 3 Pro. While most Flash type models are explicitly optimized for short, quick tasks like generating code, Google claims Gemini 3 Flash's performance "in reasoning, tool use and multimodal capabilities is ideal for developers looking to do more complex video analysis, data extraction and visual Q&A, which means it can enable more intelligent applications -- like in-game assistants or A/B test experiments -- that demand both quick answers and deep reasoning." First impressions from early users So far, early users have been largely impressed with the model, particularly its benchmark performance. What It Means for Enterprise AI Usage With Gemini 3 Flash now serving as the default engine across Google Search and the Gemini app, we are witnessing the "Flash-ification" of frontier intelligence. By making Pro-level reasoning the new baseline, Google is setting a trap for slower incumbents. The integration into platforms like Google Antigravity suggests that Google isn't just selling a model; it's selling the infrastructure for the autonomous enterprise. As developers hit the ground running with 3x faster speeds and a 90% discount on context caching, the "Gemini-first" strategy becomes a compelling financial argument. In the high-velocity race for AI dominance, Gemini 3 Flash may be the model that finally turns "vibe coding" from an experimental hobby into a production-ready reality.
[14]
Google launches Gemini 3 Flash -- and claims it's as fast as using traditional Search
In the US, Gemini 3 Pro and Nano Banana Pro are also available in AI Mode Google just launched Gemini 3 Flash, a new AI model the company says "brings the incredible reasoning of the Gemini 3 model at the speed you expect of Search." Gemini 3 Flash is available worldwide as part of Google's AI Mode, and if it lives up to the claim of being as fast as a normal person using Search, then it could be the system seller for AI search. Google says, "Gemini 3 Flash's strong performance in reasoning, tool use, and multimodal capabilities enable AI Mode to tackle your most complicated questions with greater precision - without compromising speed." And considering that up until this point, AI Mode feels more like a gimmick rather than the principal way to use Google Search, maybe Gemini 3 Flash can change the narrative. Gemini 3 Flash is completely free to use as part of AI Mode; simply navigate to the AI-powered search engine directly from Google.com. Google claims this upgraded model makes AI Mode "an all-around more powerful tool" capable of "understanding your needs, so you can ask more nuanced questions," and then, considering the constraints, "provide a thoughtful, well-formatted response." After you've got your speedy AI-powered search results, you can then explore further by clicking through sources and useful links from across the web. Users in the US are also getting access to a wider rollout of Gemini 3 Pro and the incredible AI image generation tool, Nano Banana Pro, from directly within AI Mode - a sign that Google really wants you to give AI search a try. Both of these models are available for free, but you'll need Google AI Pro or Ultra to get higher usage limits and make the most of what Gemini has to offer. I'll be totally honest, even as TechRadar's Senior AI Writer, I've had no urge to use Google's AI Mode in my day-to-day life. That said, Google is promising an AI search engine here that offers supremely fast results similar to those we've come to expect in the decades the company's search engine has ruled the internet. If this is actually true (I've yet to try Gemini 3 Flash), then AI Mode might genuinely replace Google Search, just like the company wants it to. I'm not totally convinced, especially considering that even AI Overviews in regular Google Search can often take time to load and are sometimes inaccurate. But, if any company is going to convince users to trust AI with search engine results, it's Google, and I'm very intrigued to see if the new AI model can live up to the bold marketing claims. Stay tuned to TechRadar for all Gemini news, and we'll be sure to put AI Mode's Gemini 3 Flash through its paces to see if it's as efficient as Google promises.
[15]
Google Gemini 3 Flash is fast, cheap and everywhere. Here's what to know
Why it matters: The AI race is quickly becoming a standoff between Google and OpenAI, with huge implications not just for artificial intelligence technology, but for the entire economy. Driving the news: Gemini 3 Flash includes Gemini 3 Pro's reasoning capabilities in a model that Google says is faster, more efficient and cheaper to run. * "This is about bringing the strength and the foundation of Gemini 3 to everyone," Tulsee Doshi, senior director and Gemini product lead, told Axios. The release is all about giving more people access to the most powerful AI tools, Doshi says. * As of Wednesday, Gemini 3 Flash will be the default model in the Gemini app, replacing Gemini 2.5 Flash for everyday tasks. * It will also be the default model for AI Mode in search, meaning everyday Google users worldwide will be exposed to it. * Salesforce, Workday and Figma are already using Gemini 3 Flash. * The launch comes less than a week after OpenAI launched GPT-5.2, and a day after OpenAI launched ChatGPT Images. Between the lines: More efficient AI models can put the power of machine learning into the hands of more people, including consumers and small businesses. * Google says Gemini 3 Flash excels at tasks like planning last-minute trips or learning complex educational concepts quickly. * Multimodal reasoning capabilities in the new model allow users to ask Gemini to watch videos, look at images, listen to audio or read text and turn those answers into content. The intrigue: Gemini 3 Flash performs better than Gemini 3 Pro on SWE-bench Verified, a benchmark for evaluating coding agent capabilities, Doshi says. * That makes the new model more attractive for business clients, an area where Anthropic's Claude has made heavy inroads and where OpenAI is trying to catch up. Catch up quick: Google released its Gemini 3 Pro model on November 18, lighting a fire under OpenAI and forcing the company into a "code red" mode to catch up. * Google entered into its own code red mode three years ago at the launch of ChatGPT. The big picture: In the increasingly binary AI race between OpenAI and Google, Google has the ubiquity edge. * OpenAI may have the first-mover advantage, but Gemini's distribution is wider within search and across Google's core apps. * Gemini's share of weekly mobile app downloads, monthly active users and its global website visits have all recently increased at a higher rate than ChatGPT's, according to a report in The Information. Yes, but: The relentless release cycle reflects the cut-throat nature of competition at the leading edge of the model race where any company can quickly go from leader to also-ran. * Google and OpenAI aren't far ahead enough to discount Anthropic, Meta, xAI, DeepSeek and a whole host of startups. What we're watching: Whether Google can maintain accuracy at scale as it inserts Gemini more aggressively across products.
[16]
Google reveals Gemini 3 Flash to speed up AI search and beefs up image generation
Gemini 3 Flash is now available globally within AI mode for Search. Google has just announced a new AI model in the Gemini 3 series, one that is focused on speedy responses. Say hello to Gemini 3 Flash, which is claimed to offer "frontier intelligence" and aims to speed up the Google Search experience for users. What's the big shift? Gemini 3 will be integrated within the AI Mode in Search. In case you missed it, AI mode is the new conversational version of Search that is now available as its own mode, alongside the vanilla "blue link" mode, image, news, videos, and shopping, among others. Recommended Videos The whole idea behind AI mode is to give answers just the way you would get them from an AI chatbot like Gemini and ChatGPT, instead of showing traditional web links. Aside from giving you the answer, AI mode also lets you ask follow-up questions in a conversational manner. With the release of Gemini 3 Flash, AI mode is getting a dual boost. It's smarter, faster, and the new default for AI Mode in Search. "Gemini 3 Flash's strong performance in reasoning, tool use and multimodal capabilities enable AI Mode to tackle your most complicated questions with greater precision - without compromising speed," says Google. Gemini 3 Flash in AI Mode is now available to all users worldwide. Interestingly, hours ahead of the launch, some keen eyes also spotted its presence in Google's Vertex AI platform for developers and the Canvas collaboration tool. The anonymous user mentioned on Reddit that it's much faster than the older Gemini 2.5 Flash model at creating a website using just a prompt. The more powerful Gemini models are expanding Aside from introducing the Gemini 3 Flash model, Google is also pushing the Gemini 3 Pro model for all users in the US. When you open AI Mode, you will be able to pick "Thinking with 3 Pro" from the model selector drop-down. Compared to Flash, this one is tailored for more complex queries that require deeper thinking, reasoning, and problem-solving chops. Google is also pushing its most powerful AI image model, Nano Banana Pro (Gemini 3 Pro Image), and integrating it within the Search experience. So, whether it is generating image or editing them with text prompts, the results are going to be faster and more accurate. "For both of these Pro models with AI creation tools, Google AI Pro and Ultra subscribers will have higher usage limits," says the company. On the other side of the competition, OpenAI introduced the GPT 5.2 model for ChatGPT and a next-gen image creation AI model a few days ago.
[17]
Google announces Gemini 3 Flash with Pro-level performance, rolling out now
Following last month's launch of Gemini 3 Pro, Google today announced Gemini 3 Flash for consumers and developers. The tagline, of sorts, for Gemini 3 Flash is "frontier intelligence built for speed at a fraction of the cost." It retains Gemini 3's complex reasoning, multimodal/vision understanding, and performance in agentic/vibe coding tasks, but adds "Flash-level latency, efficiency and cost." The Flash model series is Google's most popular offering. It unsurprisingly surpasses 2.5 Flash across the board and "significantly" outperforms 2.5 Pro across several benchmarks. 3 Flash is comparable to 3 Pro, while even beating it in some areas, including MMMU Pro, Toolathlon, and MPC Atlas. Google touts how 3 Flash "outperforms 2.5 Pro while being 3x faster at a fraction of the cost." This strong reasoning, tool use, and multimodal capabilities can translate to "more complex video analysis, data extraction and visual Q&A" for third-party developers building customer support agents or in-game assistants. ...priced at $0.50/1M input tokens and $3/1M output tokens (audio input remains at $1/1M input tokens). Gemini 3 Flash is rolling out to the Gemini app now and replaces 2.5 Flash as the default model. It's billed as a "major upgrade to your everyday AI" for delivering "smarts and speed." For example: ...you can quickly build fun, useful apps from scratch without prior coding knowledge. Just ask Gemini to help you iterate on an idea. You can dictate stream-of-consciousness thoughts on-the-go and turn those into a prototype. Notably, Gemini 3 Flash will be available in the model picker as two options: "Fast" for quick answers and "Thinking" for complex problems. Gemini 3 Pro will appear as "Pro" for advanced math and code prompts. It's also rolling out globally as the default model in AI Mode. Gemini 3 Flash lets you ask more nuanced questions. Meanwhile, Google is making Gemini 3 Pro with its generative UI and Nano Banana Pro available for everyone in the US. ... excels at grasping the nuance of your question to serve thoughtful, comprehensive responses that are visually digestible - pulling real-time local information and helpful links from across the web. For developers, Gemini 3 Flash is available in preview via AI Studio, Google Antigravity, Gemini CLI, and Android Studio. It joins Gemini 3 Pro and Gemini 3 Deep Think from earlier this month.
[18]
Gemini 3 Flash is rolling out globally in Google Search
This content is generated by Google AI. Generative AI is experimental Today we introduced Gemini 3 Flash, our latest model with frontier intelligence built for speed, and we're rolling it out globally in Search in AI Mode. Gemini 3 Flash brings the incredible reasoning of our Gemini 3 model at the speed you expect of Search. We're also expanding access to our Pro models in the U.S., which include powerful AI creation tools to help you visualize complex topics and deepen your understanding. Starting today, we're rolling out Gemini 3 Flash as the default model for AI Mode globally. Gemini 3 Flash's strong performance in reasoning, tool use and multimodal capabilities enable AI Mode to tackle your most complicated questions with greater precision -- without compromising speed. With this upgrade, AI Mode becomes an all-around more powerful tool. It's better at understanding your needs, so you can ask more nuanced questions and it will consider each of your constraints to provide a thoughtful, well-formatted response. And as always, you'll have access to real-time information and useful links from across the web, so you can explore further and take action.
[19]
Gemini 3 Flash launches worldwide -- here's everything to know about Google's smartest model yet
Google's Gemini just got a major speed upgrade. Gemini 3 Flash, the company's fastest AI model yet, rolled out worldwide today (December 17). But Google is betting that raw speed isn't enough anymore -- modern AI has to think well and respond instantly. As the newest model in the Gemini 3 family, Flash is designed to blur the line between lightweight assistants and high-end reasoning models. Let's just say, based on these benchmarks, OpenAI might be stuck in "Code Red" mode because this is Google's attempt to make advanced AI feel effortless. Here's what Gemini 3 Flash actually does -- and why it's a bigger deal than it sounds. Gemini 3 Flash is a new AI model built to deliver high-level reasoning at Flash-tier speeds. Google describes it as combining "frontier intelligence" with low latency, making it suitable for real-time use across apps, tools and developer workflows. This matters because users traditionally had to choose between fast models that respond quickly but reason shallowly and smart models that take longer to respond. Gemini 3 Flash aims to eliminate that trade-off. While previous "Flash" models prioritized speed over depth, Gemini 3 Flash changes that. According to Google, the model delivers Pro-level reasoning performance while maintaining faster inference times and lower compute costs. That makes it more efficient than earlier Pro models, and significantly more capable than earlier Flash versions. In practical terms, that means users will get faster answers to their complex questions and better step-by-step reasoning with more consistent performance across longer prompts. This is especially noticeable in tasks like planning, summarization and coding, where response time usually slows things down. Like other Gemini 3 models, Flash is natively multimodal. It can work with text, images, audio and video. That means you can upload a photo, ask a question about it and follow up with a related task -- without switching tools or models. Google is positioning Gemini 3 Flash as an everyday assistant that understands context across formats, not just words on a screen. And here's the kicker: you may already be using the latest Gemini Flash model without realizing it. Google says Gemini 3 Flash is rolling out across the Gemini app, where it becomes the default model for many users, and AI Mode in Google Search, powering faster, more nuanced answers. It's also a part of developer platforms including Gemini API, AI Studio, Vertex AI, Android Studio and the Gemini CLI. This model is clearly designed for everyday users who want fast, thoughtful answers, developers building interactive agent-style apps and power users who need quick iteration without sacrificing reasoning quality. Gemini 3 Flash is also especially well-suited for real-time workflows such as live coding assistance, document analysis or multimodal Q&A -- where slow responses break momentum. If it works as promised, this model could become the default experience for millions of users.
[20]
Everything to know about Google Gemini's most recent AI updates
Vibe coding, faster AI models, and real-time translation are some of the new updates from Google in Gemini. Just ahead of the holiday season, Google has rolled out a fresh set of updates to Gemini, its artificial intelligence (AI) -powered assistant, underscoring the tech giant's push to keep pace in a fast-moving AI race. The new features range from vibe coding tools that enable users to build apps without having to write code to faster models and more conversational search and translation capabilities. Euronews Next takes a look at the key updates. It's now possible to build an AI-powered app directly on Gemini. Google said this week it's integrating Opal, its tool for building mini AI apps known as Gems, directly into its web browser. The tool, introduced in July, develops the blueprint for a multi-step app that works by chaining together prompts, AI model calls, and other tools without a line of code, the company said. If users want to tweak the prompt they used to build their app, add a new feature or tool, Google said that's all possible by just describing the change they want to make. Now, users have a visual editor integrated on Gemini that notes the steps and changes made to the apps. It also lets users preview their apps before sharing them with others who have Google accounts. The visual editor also lets users upload images or video links that Opal can use as a reference when building an app. Opal is one example of a vibe coding tool, an emerging field where AI can generate apps from plain language prompts instead of using code. Some AI companies, such as OpenAI and Anthropic, released their own versions earlier this year, while Swedish vibe coding startup Lovable raised one of Europe's biggest early-stage funding of $200 million (€170 million) in its first year of operations. A Gemini model that delivers smarts and speed is Google's tagline for Gemini 3 Flash. The company said the new Flash model has PhD-level reasoning skills paired with intelligence that keeps up "with the speed of your thoughts". It has two modes: a "fast" mode to answer questions quickly and a "thinking" mode that will solve complex problems. Gemini 3 Flash is also much better than older models at understanding videos and images, and potentially turning that content into a "helpful and actionable plan" in a few seconds. The company said it is also better than older models at understanding the nuances of a user question and can provide "comprehensive responses that are visually digestible," by using real-time information from across the internet. The release of Gemini 3 Flash comes a few weeks after Google released Gemini 3 Pro, the company's "most powerful" model for vibe coding and autonomous AI agents, and Gemini 3 Deep Think, a mode that tackles complex math, science, and logic problems. Google says Gemini 3 Flash's performance "rivals larger frontier models" on the industry's benchmark tests like the GPQA Diamond, a series of complex questions in biology, physics, and chemistry, and outperforms Gemini 2.5 Flash, an earlier Flash model releasedin June. Google also made it easier to chat directly with its Search and Translation features for users in the United States. Improvements to a Live With Search feature mean that users can now have a back-and-forth voice conversation to get real-time help from the search engine. The AI-powered voice that answers the query will be "more fluid and expressive than ever before," the company said in its release. Last week, Google announced that Gemini is now integrated into the company's search and translate apps to give users "much smarter, more natural and accurate text translations" while using search. The company also released a beta version of live translation that will bring "real-time, natural-sounding" translations through headlines. One of the improvements to Translate is for it to understand idioms or phrases such as "steal my thunder," and give a more natural translation for that instead of doing it word-by-word. It will currently work for users in the United States and India when translating English to over 20 languages, including Spanish, Hindi, Japanese and German.
[21]
Gemini 3 Flash Outperforms Gemini 3 Pro and GPT 5.2 In These Key Benchmarks
Gemini 3 Flash is already live for Gemini app users and in Google Search's AI Mode. The AI wars continue to heat up. Just weeks after OpenAI declared a "code red" in its race against Google, the latter released its latest lightweight model: Gemini 3 Flash. This particular Flash is the latest in Google's Gemini 3 family, which started with Gemini 3 Pro, and Gemini 3 Deep Think. But while this latest model is meant to be a lighter, less expensive variant of the existing Gemini 3 models, Gemini 3 Flash is actually quite powerful in its own right. In fact, it beats out both Gemini 3 Pro and OpenAI's GPT-5.2 models in some benchmarks. Lightweight models are typically meant for more basic queries, for lower-budget requests, or to be run on lower-powered hardware. That means they're often faster than more powerful models that take longer to process, but can do more. According to Google, Gemini 3 Flash combines the best of both those worlds, producing a model with Gemini 3's "Pro-grade reasoning," with "Flash-level latency, efficiency, and cost." While that likely matters most to developers, general users should also notice the improvements, as Gemini 3 Flash is now the default for both Gemini (the chatbot) and AI Mode, Google's AI-powered search. You can see these improvements in Google's reported benchmarking stats for Gemini 3 Flash. In Humanity's Last Exam, an academic reasoning benchmark that tests LLMs on 2,500 questions across over 100 subjects, Gemini 3 Flash scored 33.7% with no tools, and 43.5% with search and code execution. Compare that to Gemini 3 Pro's 37.5% and 45.8% scores, respectively, or OpenAI's GPT-5.2's scores of 34.5% and 45.5%. In MMMU-Pro, a benchmark that test a model's multimodal understanding and reasoning, Gemini 3 Flash got the top score (81.2%), compared to Gemini 3 Pro (81%) and GPT-5.2 (79.5). In fact, across the 21 benchmarking tests Google highlights in its announcement, Gemini 3 Flash has the top score in three: MMMU-Pro (tied with Gemini 3 Pro), Toolathlon, and MMMLU. Gemini 3 Pro still takes the number one spot on the most tests here (14), and GPT-5.2 topped eight tests, but Gemini 3 Flash is holding its own. Google notes that Gemini 3 Flash also outperforms both Gemini 3 Pro and the entire 2.5 series in the SWE-bench Verified benchmark, which tests the model's coding agent capabilities. Gemini 3 Flash scored a 78%, while Gemini 3 Pro scored 76.2%, Gemini 2.5 Flash scored 60.4%, and Gemini 2.5 Pro scored 59.6%. (Note that GPT-5.2 scored the best of the models Google mentions in this announcement.) It's a close race, especially when you consider this is a lightweight model scoring alongside these company's flagship models. That might present an interesting dilemma for developers who pay to use AI models in their programs. Gemini 3 Flash costs $0.50 per every million input tokens (what you ask the model to do), and $3.00 per every million output tokens (the result the models returns from your prompt). Compare that to Gemini 3 Pro, which costs $2.00 per every million input tokens, and $12.00 per every million output tokens, or GPT-5.2's $3.00 and $15.00 costs, respectively. For what it's worth, it's not as cheap as Gemini 2.5 Flash ($0.30 and $2.50), or Grok 4.1 Fast for that matter ($0.20 and $0.50), but it does outperform these models in Google's reported benchmarks. Google notes that Gemini 3 Flash uses 30% fewer tokens on average than 2.5 Pro, which will save on cost, while also being three times faster. If you're someone who needs LLMs like Gemini 3 Flash to power your products, but you don't want to pay the higher costs associated with more powerful models, I could image this latest lightweight model looking appealing from a financial perspective. Most of us using AI aren't doing so as developers who need to worry about API pricing. The majority of Gemini users are likely experiencing the model through Google's consumer products, like Search, Workspace, and the Gemini app. Starting today, Gemini 3 Flash is the default model in the Gemini app. Google says it can handle many tasks "in just a few seconds." That might include asking Gemini for tips on improving your golf swing based on a video of yourself, or uploading a speech on a given historical topic and requesting any facts you might have missed. You could also ask the bot to code you a functioning app from a series of thoughts. You'll also experience Gemini 3 Flash in Google Search's AI Mode. Google says the new model is better at "parsing the nuances of your question," and thinks through each part of your request. AI Mode tries to return a more complete search result by scanning hundreds of sites at once, and putting together a summary with sources for your answer. We'll have to see if Gemini 3 Flash improves on previous iterations of AI Mode. I'm someone who still doesn't find much use for generative AI products in their day-to-day lives, and I'm not entirely sure Gemini 3 Flash is going to change that for me. However, the balance of performance gains with the cost to process that power is interesting, and I'm particularly intrigued to see how OpenAI responds. Gemini 3 Flash is available to all users starting today. In addition to general users in Gemini and AI Mode, developers will find it in the Gemini API in Google AI Studio, Gemini CLI, and Google Antigravity, the company's new agentic development platform. Enterprise users can use it in Vertex AI and Gemini Enterprise.
[22]
Google Launches Gemini 3 Flash, Promises Faster Performance and Lower Costs | AIM
Gemini 3 Flash is priced at $0.50 per million input tokens and $3 per million output tokens. Google has rolled out Gemini 3 Flash, a new AI model for developers that the company claims delivers faster performance and lowers costs while retaining advanced reasoning and multimodal capabilities. The company is making the frontier intelligence model accessible through the Gemini API via Google AI Studio, Gemini CLI, Android Studio, agentic development platform Google Antigravity, and for enterprise customers through Vertex AI. Gemini 3 Flash is priced at $0.50 per million input tokens and $3 per million output tokens, with additional cost reductions through context caching and batch processing. "Today we're introducing Gemini 3 Flash, our latest model with frontier intelligence built for speed at a fraction of the cost," said Logan Kilpatrick, group product manager at Google DeepMind. According to Google, Gemini 3 Flash builds on the capabilities of Gemini 3 Pro and outperforms Gemini 2.5 Pro across several benchmarks, while operating up to three times faster. It stated that the model supports multimodal reasoning, coding, agentic workflows, and visual understanding, including code execution for tasks such as counting, zooming, and editing visual inputs. Google said the Flash series remains its most widely used model family, processing trillions of tokens across hundreds of thousands of applications. With Gemini 3 Flash, the company aims to support large-scale production use cases that require lower latency and higher rate limits. Early users have already integrated the model into products spanning software development, gaming, document analysis, and deepfake detection. Google also recently introduced CC, an experimental AI productivity agent developed by Google Labs, to help users manage daily tasks and organise their workday more efficiently.
[23]
Google's Gemini 3 Flash makes a big splash with faster responsiveness and superior reasoning - SiliconANGLE
Google's Gemini 3 Flash makes a big splash with faster responsiveness and superior reasoning Google LLC is building on the successful launch of its all-powerful Gemini 3 Pro large language model with the debut of a more lightweight and streamlined version called Gemini 3 Flash. It's the successor to Gemini 2.5 Flash, designed for applications that require lower latency and costs, and it's being rolled out across multiple platforms starting today. Gemini 3 Flash is built on the same foundation as Gemini 3 Pro, which delivers industry-leading performance in terms of complex reasoning, multimodal and vision-based understanding, agentic artificial intelligence and coding-based tasks. Where it differs is that it's a more streamlined and efficient version of that model, sacrificing a little performance in order to reduce latency and reduce the costs of AI processing. Nonetheless, Google said Gemini 3 Flash is still one of its most impressive models for agentic workflows and can power AI agents at less than a quarter of the cost of Gemini 3 Pro, while allowing higher rate limits. More importantly, Gemini 3 Flash surpasses the performance of Gemini 2.5 Flash on numerous benchmarks, Google said, with faster time-to-first-token meaning it's even more responsive. Gemini 3 Flash is being made available now to consumers in the Gemini app, where it replaces 2.5 Flash, while developers will be able to access it via platforms including Google AI Studio, Gemini CLI, Vertex AI and Google Antigravity, a new agentic-led application development environment that debuted last month. In the Gemini app, Gemini 3 Flash becomes the new default model instead of 2.5 Flash, which means that every single user globally will benefit from its superior performance free of charge. Google Labs and Google Gemini Vice President Josh Woodward said the improved multimodal reasoning capabilities in Gemini 3 Flash mean that the app can now help people to see, hear and understand any type of information much faster than before. "For example, you can ask Gemini to understand your videos and images and turn that content into a helpful and actionable plan in just a few seconds," he said. Users can also ask Gemini to create new applications for them from scratch, even if they don't have any coding knowledge at all. It truly democratizes application development, Woodward said. Someone can just ask Gemini to help them iterate on an idea, dictate their stream-of-consciousness thoughts and transform their vision into a working prototype right there on their laptop or smartphone, he said. In addition, Gemini 3 Flash is being made the default model for Google Search's AI Mode, which responds to user's searches with detailed, AI-generated summaries to help people discover what they need to know faster. Users can expect faster, more accurate summaries, Woodward said. He explained that Gemini 3 Flash excels at understanding the nuances of user's questions, and that this allows it to generate more thoughtful and comprehensive responses based on real-time information rather than stale content. "The result effectively combines research with immediate action: you get an intelligently organized breakdown alongside specific recommendations - at the speed of Search," he promised. Developers will also benefit from enhanced performance, with Gemini 3 Flash striking the perfect balance between reasoning and speed for agentic coding tasks and responsive, interactive applications. It's available now across all of Google's major development platforms, meaning developers and their applications will be able to leverage its near real-time multimodal processing capabilities. These span complex video analysis, data extraction and visual questions and answers, and enable Gemini 3 Flash to analyze thousands of documents or video archives and generate the required insights as quick as a flash, said Gemini Senior Director of Product Management Tulsee Doshi. She explained that Gemini 3 Flash has been designed to eliminate almost completely the lag that's typically associated with larger models. That ensures split-second responsiveness for customer support agents, in-game assistants and other applications where speed is of the essence, she said. In terms of its performance, Google has tested Gemini 3 Flash across a number of popular benchmarks. It demonstrated best-in-class scores on PhD-level reasoning and knowledge benchmarks like GPQA Diamond (90.4%) and Humanity's Last Exam (33.7% without tools), rivaling many much larger frontier models. It also showed leading efficiency, outperforming Gemini 2.5 Pro by generating responses three-times faster at a fraction of the cost, Google said. Developers will also find that Gemini 3 Flash is much more cost-effective than either Gemini 3 Pro or 2.5 Flash, Doshi said. For instance, in the Gemini application programming interface and Vertex AI platforms, the model is priced at just 50 cents per 1 million input tokens and $3 per 1 million output tokens. It also comes with standard context caching, which enables cost reductions of up to 90% in applications with repeated token use above certain thresholds.
[24]
Gemini 3 Flash: frontier intelligence built for speed
This content is generated by Google AI. Generative AI is experimental Today, we're expanding the Gemini 3 model family with the release of Gemini 3 Flash, which offers frontier intelligence built for speed at a fraction of the cost. With this release, we're making Gemini 3's next-generation intelligence accessible to everyone across Google products. Last month, we kicked off Gemini 3 with Gemini 3 Pro and Gemini 3 Deep Think mode, and the response has been incredible. Since launch day, we have been processing over 1T tokens per day on our API. We've seen you use Gemini 3 to vibe code simulations to learn about complex topics, build and design interactive games and understand all types of multimodal content. With Gemini 3, we introduced frontier performance across complex reasoning, multimodal and vision understanding and agentic and vibe coding tasks. Gemini 3 Flash retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost. It not only enables everyday tasks with improved reasoning, but also is our most impressive model for agentic workflows. Starting today, Gemini 3 Flash is rolling out to millions of people globally:
[25]
Gemini 3 Flash is a game changer and here's why
Google began rolling out Gemini 3 Flash, a more efficient version of its latest AI model, approximately one month after launching Gemini 3 Pro in November. The new model delivers pro-grade reasoning performance comparable to the flagship at reduced cost, with benchmarks showing it surpasses prior generations including Gemini 2.5 Pro and competes directly with OpenAI's GPT-5.2. The rollout positions Gemini 3 Flash for widespread everyday applications due to its balance of capability and efficiency. Google states that this system maintains high-level reasoning while operating at a fraction of the computational expense of the Pro variant, enabling broader accessibility without performance trade-offs in standard tasks. Performance evaluations confirm Gemini 3 Flash exceeds Google's previous models across multiple metrics. It demonstrates substantial improvements over Gemini 2.5 Pro, reflecting advancements in model architecture and training processes that enhance speed and accuracy simultaneously. In the Humanity's Last Exam benchmark suite, which tests advanced reasoning under constrained conditions, Gemini 3 Flash achieved a score less than one percentage point below GPT-5.2. This comparison occurred without access to external tools such as web search, isolating the models' intrinsic capabilities. Gemini 3 Flash also outperformed GPT-5.2 in select evaluations focused on multimodal processing. The MMMU-Pro benchmark, designed to assess a model's understanding and reasoning across text, images, and other data types, recorded 81.2 percent for Gemini 3 Flash against 79.5 percent for GPT-5.2. Google integrates Gemini 3 Flash as the default model in the Gemini App and AI Mode within Search. This deployment extends to all users globally. "That means all of our Gemini users globally will get access to the Gemini 3 experience at no cost, giving their everyday tasks a major upgrade," Google explains. Within AI Mode, US users gain direct access to the Nano Banana Pro image generator via the chatbot interface. The process involves selecting "Thinking with 3 Pro" from the model picker, followed by "Create Images Pro."
[26]
Gemini 3 Flash Arrives as Google's Latest High-Speed, Low-Cost AI Model
Enterprises can access the model via Vertex AI and Gemini Enterprise Google released Gemini 3 Flash on Wednesday as the latest entrant in the Gemini 3 series. The artificial intelligence (AI) model joins Gemini 3 Pro and Gemini 3 Deep Think, and brings speed, efficiency, and lower token cost for users and developers. Arriving a month after the release of the previous two models, Google says the Flash variant is powerful enough to outperform 3 Pro in coding-related tasks. It is also said to be more performant compared to the entire Gemini 2.5 series.
[27]
Build with Gemini 3 Flash, frontier intelligence that scales with you
Today we're introducing Gemini 3 Flash, our latest model with frontier intelligence built for speed at a fraction of the cost. Building on 3 Pro's strong multimodal, coding and agentic features, 3 Flash offers powerful performance at less than a quarter the cost of 3 Pro, along with higher rate limits. The new 3 Flash model surpasses 2.5 Pro across many benchmarks while delivering faster speeds. It also features our most advanced visual and spatial reasoning and now offers code execution to zoom, count and edit visual inputs. Flash remains our most popular version, with 2 and 2.5 Flash processing trillions of tokens across hundreds of thousands of apps built by millions of developers. Our Flash models are truly built for developers, and with 3 Flash, you no longer need to compromise between speed and intelligence. Gemini 3 Flash is rolling out to developers in the Gemini API via Google AI Studio, Google Antigravity, Gemini CLI, Android Studio and to enterprises via Vertex AI. Gemini 3 Flash delivers frontier-class performance on PhD-level reasoning and knowledge benchmarks like GPQA Diamond (90.4%) and Humanity's Last Exam (33.7% without tools), rivaling much larger frontier models.
[28]
Google Introduces Gemini 3 Flash A Faster More Efficient AI Model for Everyday and Enterprise Use
Google has unveiled Gemini 3 Flash, a groundbreaking AI model engineered for speed, cost-effectiveness and intelligent reasoning. Positioned as the next evolution in the Gemini AI model family, this release is designed to deliver rapid responses, handle complex queries and scale across consumer and enterprise environments. Gemini 3 Flash now powers core experiences within the Gemini app and Google AI Mode in Search while offering developers robust multimodal and coding capabilities through APIs and cloud tools. Google has expanded its AI ecosystem with the introduction of Gemini 3 Flash, a model engineered to redefine how users and developers interact with artificial intelligence. Built on the latest advancements in the Gemini 3 family this release prioritizes rapid responsiveness and cost efficiency without compromising on the depth and quality of reasoning. The strategic positioning of Gemini 3 Flash reflects Google's commitment to scalable intelligence that meets the diverse needs of everyday users, developers and enterprises. At its core Gemini 3 Flash is designed around the principle that powerful AI does not have to be slow or prohibitively expensive. Traditional trade offs in AI have forced users to choose between large highly capable models that are slow and costly or lean models that respond quickly but lack reasoning depth. Gemini 3 Flash eliminates this compromise by combining advanced cognitive capabilities with flash level latency and affordability. It delivers performance that rivals larger models while significantly reducing the cost per token making it a practical choice for high frequency use cases and workflows. The model's speed improvements are substantial. Compared to earlier generations Gemini 3 Flash operates several times faster on industry benchmarks allowing it to complete complex tasks in seconds while using fewer computational resources. This performance uplift enables real time interactive experiences across platforms and supports both text and multimodal inputs including images, video and audio. Users can now receive comprehensive responses that integrate across media types facilitating richer and more intuitive interactions. One of the most visible shifts with the Gemini 3 Flash rollout is its adoption as the default AI model within the Gemini app and Google's AI Mode in Search. Making Gemini 3 Flash the standard model in these environments ensures that everyday users benefit from faster intelligent assistance for a range of needs from answering complex queries to planning tasks and generating actionable insights. In Search AI Mode the model augments traditional search results with intelligently formatted responses synthesizing information and supporting users with structured insights and recommendations. For developers and technical teams, the capabilities of Gemini 3 Flash extend far beyond simple conversational tasks. The model is now available through key tools and platforms including the Gemini API Google AI Studio Google Antigravity Gemini CLI, Android Studio Vertex AI and Gemini Enterprise. These integrations allow developers to embed powerful reasoning and multimodal understanding into applications workflows, and services while maintaining low latency and operational cost. Gemini 3 Flash delivers notable performance on reasoning and coding tasks often associated with more complex AI workloads. Benchmarks show it achieves high marks in reasoning and coding evaluations, outperforming older model generations and approaching the capabilities of larger models at a fraction of the cost. This enables rapid prototyping, iterative development and real time agentic workflows while preserving quality of output. In agentic coding contexts the model can generate working code, solve intricate problems and adapt to dynamic requests with efficiency making it useful for both exploratory projects and production environments. Enterprise customers will find additional value in the model's balance of intelligence and performance. Companies that handle large scale data analysis or need robust AI reasoning for document processing and strategic insights can deploy Gemini 3 Flash through Vertex AI and Gemini Enterprise solutions. Its ability to handle multimodal inputs at speed makes it particularly suited for complex video analysis or extracting insights from unstructured information. Beyond efficiency performance and cost advantages Gemini 3 Flash marks a broader shift toward accessible frontier intelligence. By democratizing fast intelligent AI capabilities Google is enabling a wider range of use cases and lowering barriers for innovation. Everyday users receive smarter assistance in consumer applications while developers and enterprises gain a flexible model that scales with their needs without imposing steep resource demands. In conclusion Gemini 3 Flash sets a new benchmark for AI models by delivering a balanced combination of speed reasoning, multimodal understanding and cost efficiency. Its role as the default model in major Google AI experiences and broad availability across developer platforms positions it as a transformative force in how AI assists people and powers applications across industries. Gemini 3 Flash represents a strategic step toward scalable intelligence that supports both immediate practical needs and forward looking innovation agendas.
[29]
Gemini 3 Flash : Fast, Multimodal & Easy on Your Budget
What if you could access an AI model that's not only faster and smarter but also dramatically cheaper than anything else on the market? Enter Gemini 3 Flash, Google's latest innovation that's turning heads across industries. With a bold promise of delivering up to three times the speed of its predecessor and slashing costs by over 75%, this model isn't just an upgrade, it's a redefinition of what's possible in AI. Imagine a tool that can analyze massive datasets, generate human-like reasoning, and even assist in creative tasks like game development, all while staying accessible to small developers and enterprises alike. The question isn't whether Gemini 3 Flash is powerful; it's whether the competition can keep up. In this in-depth coverage, we'll explore why Gemini 3 Flash is being hailed as a fantastic option, from its new cost-efficiency measures like Context Caching to its advanced multimodal capabilities that tackle everything from deepfake detection to enterprise-level data analysis. But is it truly the most powerful, fastest, and cheapest model ever created? Or are there trade-offs hidden beneath its impressive specs? Whether you're a developer, a business leader, or simply an AI enthusiast, this exploration will uncover how Gemini 3 Flash is reshaping the AI landscape, and what it means for the future of innovation. Sometimes, the most innovative ideas come at the most unexpected price. Gemini 3 Flash establishes a new benchmark for cost efficiency in AI technology. With pricing set at just $0.50 per 1 million input tokens and $3 per 1 million output tokens, it is less than a quarter of the cost of its predecessor, Gemini 3 Pro. This significant reduction in pricing makes it accessible to a broader audience, from small-scale developers to large enterprises. The model achieves its affordability through innovative features that optimize resource usage: These cost-saving measures ensure that Gemini 3 Flash is not only budget-friendly but also a practical choice for businesses aiming to maximize value while maintaining high-quality outputs. Gemini 3 Flash delivers a significant leap in performance, operating up to three times faster than its predecessor, Gemini 2.5 Pro. Its advanced reasoning capabilities and multimodal features enable it to tackle complex tasks with precision and efficiency. This combination of speed and functionality makes it a versatile tool for a wide range of applications. Key performance highlights include: These features ensure that Gemini 3.0 Flash can address both technical and creative challenges, making it a reliable choice for industries requiring high-speed, accurate AI solutions. Dive deeper into Google Gemini Flash with other articles and guides we have written below. The versatility of Gemini 3.0 Flash is evident in its wide-ranging applications, which span multiple industries and use cases. Its adaptability allows it to cater to the unique needs of developers, enterprises, and individuals alike. Some of its standout applications include: This broad range of applications underscores the model's ability to meet diverse demands, making it a valuable asset across various fields. Google has prioritized accessibility in the design of Gemini 3.0 Flash, making sure that it is available to a wide audience. The model can be accessed through multiple platforms, making it easy for users to integrate its capabilities into their workflows. Access options include: For those seeking free options, platforms like Kilo Code and Antigravity IDE offer opportunities to experiment with Gemini 3.0 Flash without financial commitment. This wide availability ensures that developers and organizations of all sizes can harness the model's power without significant barriers. Gemini 3.0 Flash features several strengths that make it a standout choice in the AI landscape: However, the model is not without limitations. While its performance is impressive, the output quality may fall slightly short of the Gemini 3 Pro in certain scenarios. Additionally, it may encounter minor functionality gaps in specific use cases, such as incomplete game mechanics or highly specialized tasks. Despite these limitations, Gemini 3.0 Flash remains a compelling option for users seeking a balance of affordability, speed, and advanced features. Its strengths far outweigh its weaknesses, making it a practical choice for a wide range of applications.
[30]
Google Says New AI Model Gemini 3 Flash Is Built for Speed | PYMNTS.com
The company rolled out Gemini 3 Flash globally to developers in the Gemini API in Google AI Studio, Gemini CLI and Google Antigravity; to enterprises in Vertex AI and Gemini Enterprise; and to everyone in the Gemini app and AI Mode in Search, according to a Wednesday blog post. "With Gemini 3, we introduced frontier performance across complex reasoning, multimodal and vision understanding, and agentic and vibe coding tasks," Tulsee Doshi, senior director, product management at Google DeepMind, wrote in the post. "Gemini 3 Flash retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost. It not only enables everyday tasks with improved reasoning, but also it our most impressive model for agentic workflows." For AI Mode in Search, Gemini 3 Flash will be the default model globally, according to another blog post. In addition, Gemini 3 Pro is now available in Search to everyone in the United States, and Nano Banana Pro (Gemini 3 Pro Image) is available to more people in the U.S. These models can be selected through "Thinking with 3 Pro" in the AI Mode drop-down menu. "Gemini 3 Flash's strong performance in reasoning, tool use and multimodal capabilities enable AI Mode to tackle your most complicated questions with greater precision -- without compromising speed," Robby Stein, vice president of product, Google Search, wrote in the post. In the Gemini app, Gemini 3 Flash will offer a choice of "Fast" or "Thinking" modes, while Gemini 3 Pro is available as "Pro" in the model picker and is recommended for advanced math and code, according to another blog post. "For too long, AI forced a big choice: big models that were slow and expensive, or high speed models that were less capable," Josh Woodward, vice president, Google Labs & Gemini, wrote in the post. "Gemini 3 Flash ends this compromise. Gemini 3 Flash delivers smarts and speed." The launch of Google 3 Flash follows the Nov. 18 introduction of the Gemini 3 AI model, which was then deployed directly into Google Search, the Gemini app and the company's enterprise cloud stack.
[31]
Google launches its faster AI model Gemini 3 Flash
Alphabet's (GOOG) (GOOGL) Google has globally launched its latest AI model called Gemini 3 Flash, making it the default model in the Gemini app, replacing 2.5 Flash. Last month, the company launched Gemini 3 with Gemini 3 Pro and Gemini 3 Gemini 3 Flash provides faster reasoning, better benchmark scores, improved multimodal abilities, uses fewer tokens, and outperforms Gemini 2.5 Pro in speed and cost efficiency. Gemini 3 Flash is priced at $0.50/1M input tokens and $3/1M output tokens, is three times faster than Gemini 2.5 Pro, and uses 30% fewer tokens on average for tasks. Gemini 3 Flash delivers performance rivaling larger frontier models such as OpenAI's newest GPT models, offering strong benchmark results and advanced multimodal reasoning capabilities.
[32]
Gemini Flash 3 : Practical Pick for Multimodal Reasoning at Speed & Low Cost
What if the future of AI wasn't about chasing headlines or showcasing flashy, overhyped features, but instead focused on solving real-world problems with precision and purpose? Enter Gemini 3 Flash, a model that flips the script on the AI race. Designed for practical deployment rather than fleeting buzz, this latest addition to Google AI's Gemini lineup prioritizes efficiency, scalability, and affordability. In an era where businesses are increasingly wary of ballooning costs and impractical solutions, Gemini 3 Flash offers a refreshing alternative: a tool built to deliver value where it matters most, on the ground, in the hands of those who need it. In this exploration, we'll uncover why Gemini 3 Flash stands out in a crowded AI landscape. You'll discover how its cost-effective design and adaptive reasoning capabilities make it a fantastic option for industries grappling with tight budgets and high operational demands. From its ability to handle multimodal reasoning to its knack for balancing performance with affordability, this model is a testament to what happens when innovation meets practicality. As we delve deeper, you might find yourself rethinking what "innovative" truly means in the context of AI. Gemini 3 Flash Overview Where Gemini 3 Flash Fits in the Lineup Gemini 3 Flash builds upon the advanced reasoning architecture of its sibling, Gemini 3 Pro, while introducing optimizations that emphasize efficiency and adaptability. A standout feature is its "thinking levels" mechanism, which dynamically adjusts the depth of reasoning based on the complexity of the task. This ensures that simpler tasks are processed with minimal computational overhead, while more complex challenges receive the necessary resources for accurate and thorough execution. By focusing on these targeted optimizations, Gemini 3 Flash fills a critical niche within the Gemini lineup. It provides robust AI capabilities for organizations that prioritize cost-efficiency and low latency without compromising on quality. This makes it particularly valuable for industries requiring scalable AI solutions but operating under tight budget constraints. Its design ensures that businesses can use advanced AI without the financial burden of premium models. Affordability Without Compromise One of the most compelling aspects of Gemini 3 Flash is its competitive pricing structure, which is designed to make advanced AI accessible to a broader range of users. Input processing costs are set at $0.50 per million tokens, while output generation is priced at $3 per million tokens. These rates are significantly lower than those of competing models, such as Gemini 3 Pro, Claude Sonicet 4.5, and GPT 5.2. This cost-effectiveness is particularly advantageous for organizations managing large-scale data processing or deploying AI across multiple workflows. By reducing token-based expenses, Gemini 3 Flash enables businesses to achieve their AI objectives while staying within budgetary constraints. It offers a solution that combines affordability with practical functionality, making sure that businesses can scale their operations without sacrificing quality. Gemini 3 Flash : Built for Deployment, Not Hype! Stay informed about the latest in Gemini Flash by exploring our other resources and articles. Performance That Delivers Despite its focus on affordability, Gemini 3 Flash delivers impressive performance across a variety of benchmarks, demonstrating its capability to handle complex reasoning tasks. It has achieved competitive results in the Arc AGI 2 and GPQA Diamond benchmarks, which evaluate advanced problem-solving and general-purpose question-answering abilities. In the realm of multimodal reasoning, Gemini 3 Flash excels in tasks requiring visual understanding, as evidenced by its strong performance in the MMU Pro and Screen Spot Pro benchmarks. These capabilities make it particularly effective for applications such as screenshot analysis, user interface reasoning, and other scenarios that require the integration of textual and visual inputs. Additionally, the model has proven its value in software engineering contexts, as demonstrated by its results in the Live Code Bench. This makes it a powerful tool for developers, offering support for tasks such as coding, debugging, and software design. Its versatility ensures that it can meet the diverse needs of professionals across various technical domains. Versatile Applications Gemini 3 Flash is engineered to support a wide range of applications, particularly in environments where cost-efficiency and operational speed are critical. Its ability to handle agentic workflows -- tasks requiring autonomous reasoning and decision-making -- makes it a versatile tool for industries such as: * Customer service: Automating responses, resolving queries, and improving user experiences. * Content moderation: Making sure compliance with guidelines by analyzing and filtering content efficiently. * Data analysis: Processing and interpreting large datasets to extract actionable insights. The model's multimodal capabilities enable it to process and interpret diverse data types, including text, images, and structured datasets. This makes it particularly well-suited for use cases like visual understanding, where it can analyze screenshots or interpret user interfaces with precision. Its low latency further enhances its suitability for real-time applications and high-throughput environments, making sure that it can meet the demands of dynamic operational settings. Strategic Value for Businesses Gemini 3 Flash represents a significant advancement in making sophisticated AI accessible and practical for widespread use. By balancing quality, cost, and latency, it offers a compelling price-to-performance ratio that aligns with the needs of various industries. Its design prioritizes scalability and efficiency, making it an ideal choice for businesses aiming to integrate AI into their operations at scale. For organizations seeking a cost-effective alternative to premium AI models, Gemini 3 Flash delivers consistent, high-quality performance without the associated expense. Its ability to handle diverse tasks -- ranging from data processing to autonomous decision-making -- ensures that it remains a valuable asset for businesses navigating the complexities of modern AI integration. Real-World Usability at Its Core Gemini 3 Flash is more than just another AI model; it is a thoughtfully engineered solution designed to address the practical challenges of real-world deployment. Whether you're managing large datasets, automating workflows, or exploring new AI-driven applications, Gemini 3 Flash provides the performance, affordability, and efficiency needed to achieve your goals. Its focus on practical deployment ensures that it remains a reliable and versatile tool for businesses striving to harness the power of AI in a cost-effective and impactful manner. Media Credit: Universe of AI
[33]
Gemini 3 Flash rolls out globally in Google Search, Gemini app and APIs
Google has expanded the Gemini 3 model family with the launch of Gemini 3 Flash, a new AI model designed to combine advanced reasoning with low latency and cost efficiency. The release follows the introduction of Gemini 3 Pro and Gemini 3 Deep Think and extends Gemini 3 capabilities across Search, consumer applications, developer tools, and enterprise services. Gemini 3 Flash is built on the same reasoning foundation as Gemini 3 Pro and supports complex reasoning, multimodal understanding, vision tasks, tool use, and agentic workflows. The model dynamically adjusts its reasoning depth based on task complexity, enabling faster responses for simpler queries and deeper processing for advanced use cases. It supports inputs across text, images, audio, and video, and adds advanced visual and spatial reasoning. The model also enables code execution for operations such as zooming, counting, and editing visual inputs. Benchmarks and efficiency Gemini 3 Flash delivers frontier-level performance across multiple reasoning and multimodal benchmarks: Compared to Gemini 2.5 Pro, Gemini 3 Flash uses around 30% fewer tokens on average on typical workloads. Based on Artificial Analysis benchmarking, it achieves up to three times faster inference while operating at a lower cost. Even at lower reasoning levels, the model often outperforms earlier versions running at higher reasoning settings. The model supports context caching, which can reduce costs by up to 90% in workloads with repeated token usage. It is also available through the Batch API, enabling up to 50% cost savings and higher rate limits for asynchronous processing. Paid API customers receive production-ready rate limits for synchronous and near real-time use cases. For developers, Gemini 3 Flash is optimized for iterative and high-frequency workflows. On SWE-bench Verified, it achieves a 78% score, outperforming the Gemini 2.5 series and Gemini 3 Pro while maintaining faster response times. The model supports use cases including agentic coding, video analysis, visual question answering, data extraction, document analysis, and near real-time reasoning. It is being used in areas such as game development, deepfake detection, and legal document analysis, where both speed and accuracy are required. For enterprises, the model supports production-scale deployments that require fast inference, consistent reasoning, and cost control. Gemini 3 Flash is rolling out globally as the default model for AI Mode in Google Search. With this update, AI Mode can better interpret nuanced queries, consider multiple constraints, and return structured responses while maintaining search-level speed. AI Mode continues to provide real-time information and links from across the web, supporting research, planning, comparisons, and learning tasks. Google is expanding access to Gemini 3 Pro in Search for users in the U.S. By selecting "Thinking with 3 Pro" in AI Mode, users can access deeper reasoning, interactive visual layouts, simulations, and AI creation tools. Access to Nano Banana Pro (Gemini 3 Pro Image) is also expanding in the U.S., enabling image generation and editing within Search. Higher usage limits apply to Google AI Pro and Ultra subscribers. Gemini 3 Flash is rolling out globally in the Gemini app as the default model. The app provides a "Fast" mode for quick responses and a "Thinking" mode for more complex problem-solving. Gemini 3 Pro remains available in the model picker for advanced math and coding tasks.
[34]
New Gemini 3 Flash Released : Faster, Smarter and Cheaper
What if the upgrade you didn't even know you needed suddenly became the one you couldn't live without? That's the story of Gemini 3 Flash, the latest AI model from Google DeepMind that's quietly redefining expectations. Positioned as a more budget-friendly sibling to Gemini 3 Pro, this model doesn't just hold its own, it challenges the status quo. With faster processing speeds, advanced reasoning capabilities, and multimodal adaptability, Gemini 3 Flash is proving that you don't need to break the bank for innovative performance. Whether you're a developer racing against deadlines or a creative professional seeking precision, this upgrade might just be the unexpected fantastic option you've been waiting for. In this breakdown, Prompt Engineering explore what makes Gemini 3 Flash so compelling, from its cost-efficiency to its ability to excel in time-sensitive tasks like coding and problem-solving. You'll discover how its customizable features and seamless integration options make it a versatile tool across industries, from software development to content creation. But is it truly the perfect balance of speed, affordability, and capability? Or does its lower price tag come with trade-offs you can't ignore? By the end, you'll have a clearer picture of whether Gemini 3 Flash is the upgrade your workflow has been missing, or the one you didn't see coming. Gemini 3 Flash redefines expectations in AI performance, particularly in domains like programming and logical reasoning. It outpaces its predecessors, including Gemini 2.5 Pro and Gemini 3 Pro, with faster processing speeds, making it an excellent choice for time-sensitive tasks. On critical benchmarks, it demonstrates exceptional performance: These results highlight Gemini 3 Flash's ability to deliver high-quality outputs while maintaining a competitive edge in speed and efficiency. Gemini 3 Flash is equipped with a range of advanced features that enhance its adaptability and usability across various industries. Key capabilities include: These features make Gemini 3 Flash a powerful and adaptable tool for professionals in fields ranging from software development to creative design. Check out more relevant guides from our extensive collection on Google Gemini 3 that you might find useful. Gemini 3 Flash is particularly effective for tasks that demand both precision and creativity. Its practical applications span multiple industries, including: For professionals seeking a cost-effective alternative to Gemini 3 Pro, Gemini 3 Flash delivers comparable performance while maintaining affordability, making it an attractive choice for budget-conscious projects. Gemini 3 Flash is designed for seamless integration into existing workflows, making sure minimal disruption during adoption. It is currently available in preview through platforms such as: The model supports API integration with tools and function-calling capabilities, allowing users to dynamically allocate reasoning power based on task complexity. Additionally, its shared API parameters with Gemini 3 Pro ensure a smooth transition for users already familiar with the Pro version, further simplifying its implementation. One of the standout features of Gemini 3 Flash is its affordability. Input token costs are set at $0.50 per million, with only a slight increase in output token costs. This pricing structure offers significant savings compared to Gemini 3 Pro, making it an ideal choice for developers and organizations aiming to optimize their budgets without compromising on performance. By delivering high-quality results at a fraction of the cost, Gemini 3 Flash enables users to achieve their goals more efficiently. While Gemini 3 Flash offers impressive capabilities, it is important to be aware of its limitations to ensure it aligns with your specific needs: Understanding these limitations can help users make informed decisions about when and how to deploy Gemini 3 Flash effectively. To fully use the strengths of Gemini 3 Flash, consider using it in tandem with Gemini 3 Pro. While Gemini 3 Pro excels in planning and orchestration, Gemini 3 Flash is optimized for implementation tasks, offering a balance of speed, cost, and capability. This complementary approach allows users to maximize efficiency and value, making sure the best possible outcomes for their projects. Gemini 3 Flash represents a significant step forward in accessible AI technology, offering a robust combination of performance, versatility, and cost-effectiveness. By understanding its features, applications, and limitations, users can unlock its full potential and achieve exceptional results across a wide range of tasks.
[35]
Google Launches Gemini 3 Flash AI: Here's What's New and How You Can Use It
has launched Gemini 3 Flash, which delivers high-level intelligence at much faster speeds and lower cost. The tech giant is making this model accessible to everyone, applicable across the company's products. In a recent blog, Saurabh Tiwary, VP & GM of Cloud AI, shared that the Gemini 3 Flash is built to be highly efficient, pushing the boundaries of quality at better price, performance, and speed. Tiwary further added, "With a near real-time response from the model, businesses can now provide more engaging experiences for their end users at production scale, without sacrificing on quality."
[36]
Google launches Gemini 3 Flash for faster AI processing By Investing.com
Investing.com -- Google has released Gemini 3 Flash, a new AI model designed to deliver high-level intelligence with faster processing speeds at lower costs. The new model, launched Wednesday, joins the Gemini 3 family that was introduced last month with Gemini 3 Pro and Gemini 3 Deep Think mode. Since the initial launch, Google has been processing over 1 trillion tokens daily through its API. Gemini 3 Flash maintains the advanced reasoning capabilities of its predecessors while offering improved speed and efficiency. The model achieves 90.4% on the GPQA Diamond benchmark and 33.7% on Humanity's Last Exam without tools, matching larger frontier models. It also scores 81.2% on MMMU Pro, comparable to Gemini 3 Pro. The new model uses 30% fewer tokens on average than Gemini 2.5 Pro while delivering higher performance. It processes information three times faster at a fraction of the cost, priced at $0.50 per million input tokens and $3 per million output tokens. For developers, Gemini 3 Flash achieves a 78% score on SWE-bench Verified, outperforming both the 2.5 series and Gemini 3 Pro. Companies including JetBrains, Bridgewater Associates, and Figma are already using the model. Gemini 3 Flash is now the default model in the Gemini app globally, replacing Gemini 2.5 Flash. It's also beginning to roll out as the default model for AI Mode in Search worldwide. The model is available through the Gemini API in Google AI Studio, Google Antigravity, Vertex AI, Gemini Enterprise, and other developer tools including Gemini CLI and Android Studio. This article was generated with the support of AI and reviewed by an editor. For more information see our T&C.
[37]
Google launches Gemini 3 Flash AI model: What's new and how to access
Google claims that Gemini 3 Flash is "our most impressive model for agentic workflows." Google has launched Gemini 3 Flash, a new AI model built to deliver high-level intelligence at much faster speeds and lower cost. It is the latest addition to the Gemini 3 family, following the recent release of Gemini 3 Pro and Gemini 3 Deep Think. The tech giant claims that the Gemini 3 Flash is "our most impressive model for agentic workflows." Google is making this model accessible to everyone across the company's products. Keep reading for all the details. Gemini 3 Flash focuses on speed without giving up intelligence. It combines the strong reasoning abilities of Gemini 3 Pro with low-latency. This allows it to answer questions quickly while still handling complex tasks such as advanced reasoning, coding, and understanding images and videos. The model performs at a high level on difficult academic and knowledge benchmarks, showing results similar to much larger AI systems. Google claims Gemini 3 Flash can adjust how much it "thinks," spending more time on difficult problems and responding faster to simple ones. Also read: Meta encourages staff to use OpenAI and Google models in AI-first workplace push: Report Gemini 3 Flash is rolling out globally. For regular users, Gemini 3 Flash is becoming the default model in the Gemini app, replacing Gemini 2.5 Flash at no cost. It is also rolling out as the default model for AI Mode in Google Search. Also read: OpenAI launches faster ChatGPT Images with GPT Image 1.5 to rival Gemini Nano Banana: How to use it Developers can access the Gemini 3 Flash model through the Gemini API in Google AI Studio, Gemini CLI, Vertex AI, Gemini Enterprise, and Google's new agentic platform, Google Antigravity. Also read: Elon Musk's X sues startup over attempt to claim Twitter trademark: Here's what happened
Share
Share
Copy Link
Google has released Gemini 3 Flash, making it the default in the Gemini app and AI Mode in Google Search globally. The new AI model combines the powerful reasoning capabilities of Gemini 3 Pro with faster performance and lower operational costs. Early benchmarks show significant improvements over its predecessor, with companies like JetBrains, Figma, and Cursor already deploying the model.
Google has launched Gemini 3 Flash, positioning the new AI model as the default in the Gemini app and AI Mode in Google Search globally
1
2
. The release arrives just one month after Gemini 3 Pro debuted and six months following the Gemini 2.5 Flash announcement, marking an accelerated push in Google's competition with OpenAI and other AI developers4
. Tulsee Doshi, Senior Director and Head of Product for Gemini Models at Google DeepMind, describes the transition as a "huge upgrade" for most users, promising faster turnaround from a latency perspective alongside more detailed, nuanced answers compared to the previous generation5
.
Source: Google
The faster and more efficient AI model is now available for developers through the Gemini API, Vertex AI, AI Studio, Google Antigravity, Gemini CLI, and Android Studio
1
4
. Companies including JetBrains, Figma, Cursor, Harvey, and Latitude are already using Gemini 3 Flash in production environments2
.Gemini 3 Flash demonstrates substantial improvements in testing, with particularly striking gains in specialized assessments. On Humanity's Last Exam, which tests advanced domain-specific knowledge, the model scored 33.7% without tool use—a threefold increase over Gemini 2.5 Flash's 11% score and competitive with GPT-5.2's 34.5%
2
. The model achieved 81.2% on the MMMU-Pro multimodality and reasoning benchmark, outperforming all competitors including Gemini 3 Pro2
.Coding capabilities show marked advancement, with Gemini 3 Flash gaining almost 20 points on the SWE-Bench Verified test compared to the 2.5 branch
1
. The model outperformed both Gemini 3 Pro and Claude Sonnet 4.5 on this agentic coding benchmark4
. In the GPQA Diamond test requiring PhD-level reasoning, Gemini 3 Flash achieved 90.4%, approaching Gemini 3 Pro's 91.9%3
. General knowledge accuracy also improved dramatically, with Simple QA Verified scores jumping from 28.1% in Gemini 2.5 Flash to 68.7% in the new model1
.Gemini 3 Flash retains the powerful reasoning capabilities and multimodal capabilities that defined Gemini 3 Pro, including the ability to process images, text, videos, and audio simultaneously
1
4
. Users can upload pickleball videos for technique analysis, submit sketches for identification, or provide audio recordings for detailed assessment2
. The model can generate app prototypes directly within the Gemini app using natural language prompts and create plans based on series of videos and images in just seconds4
5
.
Source: TechRadar
For developers, pricing is set at $0.50 per 1 million input tokens and $3.00 per 1 million output tokens—a modest increase from Gemini 2.5 Flash's $0.30 and $2.50 respectively, but substantially lower than Gemini 3 Pro's $2 and $12 rates
1
2
. Google claims Gemini 3 Flash outperforms Gemini 2.5 Pro while operating three times faster and using 30% fewer tokens on average for thinking tasks, potentially offsetting the higher per-token cost1
2
. Doshi positioned the model as "more of your workhorse model" suitable for bulk tasks and quick, repeatable workflows like customer support agents or in-game assistants2
4
.Related Stories
The rapid deployment reflects intensifying competition in the AI sector. Since releasing Gemini 3, Google has processed over 1 trillion tokens per day through its API
2
. Reports indicate OpenAI CEO Sam Altman sent an internal "Code Red" memo after ChatGPT traffic declined as Google's market share rose, prompting OpenAI to release GPT-5.2 and a new image generation model2
. OpenAI claims ChatGPT message volume has grown 8x since November 2024 [2](https://techcrunch.com/2025/12/17/google-launches-gemini-3-flash-makes-it-the-default-model-in-the- Gemini-app/).Google's decision to immediately embed Gemini 3 Flash into its cornerstone products signals strong internal confidence, contrasting with the cautious, gradual rollouts that characterized earlier phases of the AI race
4
. Gemini 3 Pro and Nano Banana Pro, Google's image model, are also expanding availability in AI Mode for all US-based users, with free access subject to unspecified limits while Pro and Ultra subscribers receive higher usage allowances1
3
. Users can still select Gemini 3 Pro from the model picker for specialized math and coding questions2
.
Source: Digital Trends
Summarized by
Navi
[2]
[4]
08 Apr 2025•Technology

18 Jun 2025•Technology

31 Jan 2025•Technology

1
Policy and Regulation

2
Technology

3
Technology
