14 Sources
[1]
Google's Gemini AI family updated with stable 2.5 Pro, super-efficient 2.5 Flash-Lite
Google has announced a big expansion of its Gemini AI model family today. After months of tweaking and tuning, the high-power Gemini 2.5 Pro is leaving preview and is ready for developers to build on. Meanwhile, Google is offering a peek at its upcoming high-efficiency model, known as Gemini 2.5 Pro Flash-Lite. Try as it might, Google can't get away from confusing model names. Google's AI aspirations have been looking up in 2025 with the debut of Gemini 2.5. These models showed a marked improvement over past versions, making Google more competitive with OpenAI and its popular GPT models. However, we've been inundated with previews and test builds as Google works toward general availability, which means a model is stable enough for long-term development work. The 2.5 Flash model left preview at I/O, but Gemini 2.5 Pro lagged behind. Today, Flash is hitting general availability with the 04-17 build. Gemini 2.5 Pro is leaving preview and also reaching general availability, and as predicted, the recently revamped 06-05 build is the winner. This version aimed to address some issues that popped up in the Google I/O build of 2.5 Pro, and it appears to have worked.
[2]
Gemini 2.5 Pro and Flash roll out to Google AI subscribers - how to access
The company also introduced Gemini 2.5 Flash-Lite, a model built for speed and cost-efficiency. Google is pushing forward rapidly with updates to Gemini 2.5, the latest addition to the company's family of AI models, which debuted earlier this year. On Tuesday, the tech giant announced that Gemini 2.5 Pro and Flash -- both of which were originally released as limited public previews -- are now generally available and stable. That means they've been deemed reliably safe for public use, a win for users. Google also unveiled a new addition to the family -- Gemini 2.5 Flash-Lite. Also: You can now generate images with ChatGPT on WhatsApp - here's how Gemini 2.5 Pro and Flash are "thinking" models designed to undergo a multi-step reasoning process when responding to user prompts. This process boosts their accuracy and performance, similar to the capabilities found in OpenAI's o1 and DeepSeek's R1. Developers can also adjust the models' "thinking budget," increasing or decreasing the degree to which they'll reason through their responses. (Disclosure: Ziff Davis, ZDNET's parent company, filed an April 2025 lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.) AI models typically undergo rigorous internal safety testing before being deployed to the public. This process has become much more critical in recent years as public and political concerns about the tendency for generative AI models to hallucinate or present false information with the veneer of fact continue to grow. That safety testing is likely why it took Google some time to launch the models into general availability. Google also recently announced that the limit of daily user prompts for Gemini 2.5 Pro would be doubled from 50 to 100, signaling a growing confidence in the model's ability to engage in longer conversations. The models are now accessible through Google AI Studio, Vertex AI, and the Gemini app. Google also introduced on Tuesday a new member of its Gemini family of models called 2.5 Flash-Lite. Available now in public preview through Google AI Studio and Vertex AI, 2.5 Flash-Lite is designed to be a more cost-effective alternative to its Pro and Flash counterparts. The model has the lowest latency of any of the Gemini models, according to Google, meaning developers won't have to sacrifice speed for cost. It's also a reasoning model; however, this capability is off by default to optimize speed. Developers have the option to turn it on by adjusting its thinking budget. Also: ChatGPT search just got smarter - but can it replace Google for you yet? The pricing for the Gemini family of models varies depending on the model and the type of content being generated--more info can be found here. The expansion of Gemini, Google's flagship generative AI model, into various models of different sizes and capabilities reflects a broader trend underway in Silicon Valley: tech developers are competing to make their tools accessible to as many people as possible. By releasing models with different capabilities and price points, Google is making it easier for customers to access Gemini regardless of their professional needs or budgets. The company has made Gemini a core part of search with AI Overviews, and the model has also been integrated into Google Meet, Docs, Sheets, and more of the company's product portfolio. Also: This new ChatGPT trick lets you download your Canvas work as a PDF in seconds - here's how OpenAI has been making a similar effort to maximize the reach of ChatGPT by, for example, integrating the chatbot with WhatsApp.
[3]
Google's Gemini 2.5 models are out of preview and available to all
It's only been a few years since OpenAI showcased how powerful AI tools can be with ChatGTP. The AI tool took the world by storm, forever altering how we think and interact with AI models. It was very clear at that point, that going forward, this would be the future. Of course, Google, MIcrosoft, and other large companies would also step up to the plate, looking to introduce AI tools of its own. Related Google Gemini: Everything you need to know about Google's next-gen multimodal AI Google Gemini is here, with a whole new approach to multimodal AI Posts Google's efforts would eventually yield Gemini, which we've seen evolve over time, offering different tiers of service. We're now getting news from Google, sharing that its Gemini 2.5 Flash and Pro models will be available to all as they hit a stable release (via 9to5Google). In addition, the brand is also debuting a new model in preview called Gemini 2.5 Flash-Lite. With so many options, it can all be a bit confusing Source: Google Now, we get it, this can all be a little complicated if you're someone that just wants to dabble with AI and get a handle on Google's tools. If you're someone that pays for the service, you probably won't get much use out of this change, since it is aimed more towards developers. As mentioned before, all developers will now have the ability to access Google's Gemini 2.5 Flash and Pro models. Google has also introduced a preview of its new Gemini 2.5 Flash-Lite model, which it claims is its fastest model yet. Developers can start working with this model in preview starting today. Now, this can all be a bit confusing, but luckily the brand has issued a chart that breaks down how each of its Gemini 2.5 models works, and what it costs when it comes to inputs and outputs. You can check out the chart which breaks everything down in the image above, or if you're someone that likes to go a bit more in depth, you can always read the huge breakdown of Google's models if you want to really do a deep dive. If not, you can always check out our quick explainer on what Gemini is and what tools it offers. For the most part, this is just one more step in a long journey for AI. For most of us, it will be interesting to see just how these tools get utilized and how helpful they can be when they are introduced to new apps and services. Google will no doubt showcase its AI features in its latest products, like the Pixel 10, which we could see release really soon.
[4]
Google launches Gemini 2.5 Pro, free app access continues
Following the first preview three months ago, Google is ready to launch Gemini 2.5 Pro and make the reasoning model generally available (GA) for consumers and developers. Gemini 2.5 Pro entered experimental testing in late March (03-25). Four days after debuting for paying subscribers and developers, Google surprisingly made it available for free users. It got a big coding upgrade in May (05-06) just before I/O, and one last update at the start of June (06-05). With the stable version today, there are "no changes from the 06-05 preview version." L-R: Latest, 06-05, 05-06, 03-25 benchmarks 2.5 Pro will lose the "preview" label in the Gemini app's model picker. It follows 2.5 Flash entering GA in the app at I/O last month. Google notes how Pro is for "Reasoning, math & code" prompts, while Flash is for "Fast all-around help." Free Gemini users will continue to get "limited access" to 2.5 Pro, with Google AI Pro subscribers getting 100 prompts per day, or "expanded access." Google AI Ultra offers the "highest access." 2.5 Flash is also now generally available and stable for developers. It's the same 05-20 model preview from I/O with updated pricing: Additionally, developers can now preview Gemini 2.5 Flash Lite for "high-volume, latency-sensitive tasks like translation and classification" where cost is a priority. Google touts "lower latency than 2.0 Flash-Lite and 2.0 Flash on a broad sample of prompts." Thinking is turned off by default, but can be enabled with set budgets. Available native tools include "Grounding with Google Search, Code Execution, and URL Context in addition to function calling." There's also multimodal input and the 1 million-token context length.
[5]
Google launches production-ready Gemini 2.5 AI models to challenge OpenAI's enterprise dominance
Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more Google moved decisively to strengthen its position in the artificial intelligence arms race Monday, declaring its most powerful Gemini 2.5 models ready for enterprise production while unveiling a new ultra-efficient variant designed to undercut competitors on cost and speed. The Alphabet subsidiary promoted two of its flagship AI models -- Gemini 2.5 Pro and Gemini 2.5 Flash -- from experimental preview status to general availability, signaling the company's confidence that the technology can handle mission-critical business applications. Google simultaneously introduced Gemini 2.5 Flash-Lite, positioning it as the most cost-effective option in its model lineup for high-volume tasks. The announcements represent Google's most assertive challenge yet to OpenAI's market leadership, offering enterprises a comprehensive suite of AI tools spanning from premium reasoning capabilities to budget-conscious automation. The move comes as businesses increasingly demand production-ready AI systems that can scale reliably across their operations. Why Google finally moved its most powerful AI models from preview to production status Google's decision to graduate these models from preview reflects mounting pressure to match OpenAI's rapid deployment of consumer and enterprise AI tools. While OpenAI has dominated headlines with ChatGPT and its GPT-4 family, Google has pursued a more cautious approach, extensively testing models before declaring them production-ready. "The momentum of the Gemini 2.5 era continues to build," wrote Jason Gelman, Director of Product Management for Vertex AI, in a blog post announcing the updates. The language suggests Google views this moment as pivotal in establishing its AI platform's credibility among enterprise buyers. The timing appears strategic. Google released these updates just weeks after OpenAI faced scrutiny over the safety and reliability of its latest models, creating an opening for Google to position itself as the more stable, enterprise-focused alternative. How Gemini's 'thinking' capabilities give enterprises more control over AI decision-making What distinguishes Google's approach is its emphasis on "reasoning" or "thinking" capabilities -- a technical architecture that allows models to process problems more deliberately before responding. Unlike traditional language models that generate responses immediately, Gemini 2.5 models can spend additional computational resources working through complex problems step-by-step. This "thinking budget" gives developers unprecedented control over AI behavior. They can instruct models to think longer for complex reasoning tasks or respond quickly for simple queries, optimizing both accuracy and cost. The feature addresses a critical enterprise need: predictable AI behavior that can be tuned for specific business requirements. Gemini 2.5 Pro, positioned as Google's most capable model, excels at complex reasoning, advanced code generation, and multimodal understanding. It can process up to one million tokens of context -- roughly equivalent to 750,000 words -- enabling it to analyze entire codebases or lengthy documents in a single session. Gemini 2.5 Flash strikes a balance between capability and efficiency, designed for high-throughput enterprise tasks like large-scale document summarization and responsive chat applications. The newly introduced Flash-Lite variant sacrifices some intelligence for dramatic cost savings, targeting use cases like classification and translation where speed and volume matter more than sophisticated reasoning. Major companies like Snap and SmartBear are already using Gemini 2.5 in mission-critical applications Several major companies have already integrated these models into production systems, suggesting Google's confidence in their stability isn't misplaced. Snap Inc. uses Gemini 2.5 Pro to power spatial intelligence features in its AR glasses, translating 2D image coordinates into 3D space for augmented reality applications. SmartBear, which provides software testing tools, leverages Gemini 2.5 Flash to translate manual test scripts into automated tests. "The ROI is multifaceted," said Fitz Nowlan, the company's VP of AI, describing how the technology accelerates testing velocity while reducing costs. Healthcare technology company Connective Health uses the models to extract vital medical information from complex free-text records -- a task requiring both accuracy and reliability given the life-or-death nature of medical data. The company's success with these applications suggests Google's models have achieved the reliability threshold necessary for regulated industries. Google's new AI pricing strategy targets both premium and budget-conscious enterprise customers Google's pricing decisions signal its determination to compete aggressively across market segments. The company raised prices for Gemini 2.5 Flash input tokens from $0.15 to $0.30 per million tokens while reducing output token costs from $3.50 to $2.50 per million tokens. This restructuring benefits applications that generate lengthy responses -- a common enterprise use case. More significantly, Google eliminated the previous distinction between "thinking" and "non-thinking" pricing that had confused developers. The simplified pricing structure removes a barrier to adoption while making cost prediction easier for enterprise buyers. Flash-Lite's introduction at $0.10 per million input tokens and $0.40 per million output tokens creates a new bottom tier designed to capture price-sensitive workloads. This pricing positions Google to compete with smaller AI providers who have gained traction by offering basic models at extremely low costs. What Google's three-tier model lineup means for the competitive AI landscape The simultaneous release of three production-ready models across different performance tiers represents a sophisticated market segmentation strategy. Google appears to be borrowing from the traditional software industry playbook: offer good, better, and best options to capture customers across budget ranges while providing upgrade paths as needs evolve. This approach contrasts sharply with OpenAI's strategy of pushing users toward its most capable (and expensive) models. Google's willingness to offer genuinely low-cost alternatives could disrupt the market's pricing dynamics, particularly for high-volume applications where cost per interaction matters more than peak performance. The technical capabilities also position Google advantageously for enterprise sales cycles. The million-token context length enables use cases -- like analyzing entire legal contracts or processing comprehensive financial reports -- that competing models cannot handle effectively. For large enterprises with complex document processing needs, this capability difference could prove decisive. How Google's enterprise-focused approach differs from OpenAI's consumer-first strategy These releases occur against the backdrop of intensifying AI competition across multiple fronts. While consumer attention focuses on chatbot interfaces, the real business value -- and revenue potential -- lies in enterprise applications that can automate complex workflows and augment human decision-making. Google's emphasis on production readiness and enterprise features suggests the company has learned from earlier AI deployment challenges. Previous Google AI launches sometimes felt premature or disconnected from real business needs. The extensive preview period for Gemini 2.5 models, combined with early enterprise partnerships, indicates a more mature approach to product development. The technical architecture choices also reflect lessons learned from the broader industry. The "thinking" capability addresses criticism that AI models make decisions too quickly, without sufficient consideration of complex factors. By making this reasoning process controllable and transparent, Google positions its models as more trustworthy for high-stakes business applications. What enterprises need to know about choosing between competing AI platforms Google's aggressive positioning of the Gemini 2.5 family sets up 2025 as a pivotal year for enterprise AI adoption. With production-ready models spanning performance and cost requirements, Google has eliminated many of the technical and economic barriers that previously limited enterprise AI deployment. The real test will come as businesses integrate these tools into critical workflows. Early enterprise adopters report promising results, but broader market validation requires months of production use across diverse industries and applications. For technical decision makers, Google's announcement creates both opportunity and complexity. The range of model options enables more precise matching of capabilities to requirements, but also demands more sophisticated evaluation and deployment strategies. Organizations must now consider not just whether to adopt AI, but which specific models and configurations best serve their unique needs. The stakes extend beyond individual company decisions. As AI becomes integral to business operations across industries, the choice of AI platform increasingly determines competitive advantage. Enterprise buyers face a critical inflection point: commit to a single AI provider's ecosystem or maintain costly multi-vendor strategies as the technology matures. Google wants to become the enterprise standard for AI -- a position that could prove extraordinarily valuable as AI adoption accelerates. The company that created the search engine now wants to create the intelligence engine that powers every business decision. After years of watching OpenAI capture headlines and market share, Google has finally stopped talking about the future of AI and started selling it.
[6]
Google Gemini's super-fast Flash-Lite 2.5 model is out now - here's why you should switch today
The new model is in preview, while Gemini 2.5 Flash and Pro are now generally available AI chatbots can respond at a pretty rapid clip at this point, but Google has a new model aimed at speeding things up even more under the right circumstances. The tech giant has unveiled the Gemini 2.5 Flash-Lite model as a preview, joining the larger Gemini family as the smaller, yet faster and more agile sibling to the Gemini 2.5 Flash and Gemini 2.5 Pro. Google is pitching Flash-Lite as ideal for tasks where milliseconds matter and budgets are limited. It's intended for tasks that may be large but relatively simple, such as bulk translation, data classification, and organizing any information. Like the other Gemini models, it can still process requests and handle images and other media, but the principal value lies in its speed, which is faster than that of the other Gemini 2.5 models. It's an update of the Gemini 2.0 Flash-Lite model. The 2.5 iteration has performed better in tests than its predecessor, especially in math, science, logic, and coding tasks. Flash-Lite is about 1.5 times faster than older models. The budgetary element also makes Flash-Lite unique. While other models may turn to more powerful, and thus more expensive, reasoning tools to answer questions, Flash-Lite doesn't always default to that approach. You can actually flip that switch on or off depending on what you're asking the model to do. And just because it can be cheaper and faster doesn't mean Flash-Lite is limited in the scale of what it can do. Its context window of one million tokens means you could ask it to translate a fairly hefty book, and it would do it all in one go. The preview release of Flash-Lite isn't Google's only AI model news. The Gemini 2.5 Flash and Pro models, which have been in preview, are now generally available. The growing catalogue of Gemini models isn't just a random attempt by Google to see what people like. The variations are tuned for specific needs, making it so Google can pitch Gemini as a whole to a lot more people and organizations, with a model to match most needs. Flash-Lite 2.5 isn't about being the smartest model, but in many cases, its speed and price make it the most appealing. You don't need tons of nuance to classify social media posts, summarize YouTube transcripts, or translate website content into a dozen languages. That's exactly where this model thrives. And while OpenAI, Anthropic, and others are releasing their own fast-and-cheap AI models, Google's advantage in integration with its other products likely helps it pull ahead in the race against its AI rivals.
[7]
Google's latest Gemini 2.5 models are its biggest response to ChatGPT yet -- and they're already live
Google brings Gemini's latest 2.5 Flash and Pro models to audiences, and makes Flash-Lite available for testing. Gemini, Google's family of AI models, are often seen as playing second fiddle to OpenAI's GPT family, which powers ChatGPT -- but they've come a long way since a bumpy launch, with Google's AI proving itself to be a great fit for those looking for anything from a speedy research aide to a solid smartphone assistant. Unlike ChatGPT, Gemini is designed to run across a multitude of services, be it drafting messages in Gmail, transcribing calls in Google Meet, or acting as a second set of eyes and ears as the AI assistant for future Android XR smart glasses. Now, Google has announced it is rolling out its latest upgrades to the models that make Gemini so powerful, pushing preview models like Gemini 2.5 Pro and Flash out of preview and into general availability -- and making a new Flash-Lite model open for testing. Here's what's new, and, more importantly, what it means for you. If you're new to Gemini, or you're just a casual AI user, then talk of models, tokens, and benchmarks is likely to cause more confusion than clarity. So let's keep it simple: Google now offers three updated Gemini 2.5 models designed to tackle different kinds of tasks; Flash, Pro, and a newly introduced Flash-Lite. For those already familiar with Google's models, the new Gemini 2.5 family is better at memory, reasoning, and responses, with improved performance and accuracy across the board. All three models support multimodal inputs (text, images, and audio), tool use (like code execution), and have a massive 1 million token context window -- ideal for handling large documents or specific and detailed prompts. Both Gemini 2.5 Flash and 2.5 Pro also feature improved native audio output for more natural-sounding conversations, and enhanced reasoning with Deep Think. For developers, Gemini 2.5 Flash now features rebalanced pricing, with slightly more expensive input tokens being met with cheaper output tokens. Gemini 2.5 Flash and 2.5 Pro can now be accessed freely on the Gemini web portal and the Gemini mobile app by selecting your model of choice from the drop-down menu at the top of the conversation window.
[8]
Google updates Gemini 2.5 LLM series with new entry-level model, pricing changes - SiliconANGLE
Google LLC today introduced a new large language model, Gemini 2.5 Flash-Lite, that can process prompts faster and more cost-efficiently than its predecessor. The algorithm is rolling out as part of a broader update to the company's flagship Gemini 2.5 LLM series. The two existing models in the lineup, Gemini 2.5 Flash and Gemini 2.5 Pro, have moved from preview to general availability. The latter algorithm also received several pricing changes. Gemini 2.5 made its original debut in March. The LLMs in the series are based on a mixture-of-experts architecture, which means that they each comprise multiple neural networks. When a user submits a prompt, Gemini 2.5 activates only one of the neural networks rather than all of them, which lowers hardware usage. The LLM series is the first that Google trained using its internally-developed TPUv5p AI chip. According to the company, the training processing involved multiple server clusters that each contained 8,960 TPUv5p chips. Google's researchers equipped the clusters with new software that can automatically mitigate some technical issues. Gemini 2.5 models are multimodal with support for up to one million tokens per prompt. Google describes the flagship algorithm in the series, Gemini 2.5 Pro, as its most capable LLM to date. During internal tests, it outperformed OpenAI's o3-mini across a range of math and coding benchmarks. Gemini 2.5 Flash, the model that moved into general availability today together with Gemini 2.5 Pro, trades off some performance for efficiency. It responds to prompts faster and incurs lower inference costs. Gemini 2.5 Flash-Lite, the new model that Google debuted today, is an even more efficient model that is positioned as the new entry-level model in the LLM series. "2.5 Flash Lite has all-around higher quality than 2.0 Flash-Lite on coding, math, science, reasoning and multimodal benchmarks," Tulsee Doshi, the senior director of product management for Gemini, detailed in a blog post. "It excels at high-volume, latency-sensitive tasks like translation and classification, with lower latency than 2.0 Flash-Lite and 2.0 Flash on a broad sample of prompts." Gemini 2.5 Flash-Lite is billed at a rate of $0.1 per 1 million input tokens when developers submit prompts that contain text, images or video. That's less than one tenth the cost of Gemini 2.5 Pro. The price per million tokens of output, in turn, is $0.4 compared with $10 for Gemini 2.5 Pro. Google is changing the pricing of its mid-range Gemini 2.5 Flash model as part of the update. The company will now charge $0.3 per million input tokens and $2.5 per 1 million output tokens compared with $.15 and $3.5, respectively, before. Additionally, there is no longer separate pricing for tokens that the model processes in "thinking mode." The mode allows the LLM to boost output quality by increasing the amount of time and compute resources that it uses to generate prompt responses.
[9]
You Can Now Use the Stable Version of Gemini 2.5 Pro and Flash AI Models
Google made the Gemini 2.5 family of artificial intelligence (AI) models generally available on Tuesday. With this, users of the chatbot can now access the stable versions of the Gemini 2.5 Pro and the Gemini 2.5 Flash models. Interestingly, the Mountain View-based tech giant has also made the Pro model available to the users on the free tier of the Gemini platform. Additionally, the company has also released a 2.5 Flash-Lite, which is claimed to be Google's fastest and most cost-efficient AI model. In a blog post, the tech giant announced the rollout of the stable versions of the Gemini 2.5 Pro and Flash models. These large language models (LLMs) were so far available to users as a preview, meaning the full range of capability could not be used. While in preview, these models also tend to be prone to errors and glitches, which are likely to be fixed with the stable version. While the Google AI Pro and Ultra users will continue to get access to the Gemini 2.5 Pro model, those on the free tier can also use it. However, the daily limit for free users is expected to be lower than the paid users. Notably, Google AI Pro users get expanded access to the model with 100 daily prompts, and Ultra users get an even higher rate limit. Notably, this version of the Pro model is similar to the one released earlier this month, and there are no notable changes. The change also means that the model selector menu on the Gemini website and app will no longer show the preview versions of these models. Those on the free tier will now see the Gemini 2.5 Flash, Gemini 2.5 Pro, and the Personalisation Preview model which can access the user's Google Search history and answer queries based on that. Notably, the tech giant has also released the Gemini 2.5 Flash-Lite model. Google says it has higher performance than the 2.0 Flash-Lite and fares better in areas such as coding, mathematics, science, reasoning, and multimodal tasks. The low-latency model is aimed at near real-time tasks such as translation and classification. It also gets the other features from the 2.5 family such as reasoning at different token budgets, connecting to Google Search and code execution tools, multimodal input support, and a context window of one million tokens. Gemini 2.5 Flash-Lite is currently available via Google AI Studio and Vertex AI. These platforms are also hosting the stable versions of 2.5 Pro and Flash. Additionally, Google is also integrating 2.5 Flash-Lite and Flash models to Search.
[10]
Google introduces stable Gemini 2.5 Flash and Pro, previews Gemini 2.5 Flash-Lite
Google DeepMind released stable versions of Gemini 2.5 Flash and Pro, highlighting top benchmark performance. Now broadly available to developers, the models support advanced reasoning, multimodal tasks, and one million-token context. A new Flash-Lite preview offers the fastest, most cost-efficient model in the Gemini 2.5 series.Google DeepMind has introduced the stable versions of its Gemini 2.5 Flash and Gemini 2.5 Pro models, its chief executive Sundar Pichai confirmed on Tuesday. ET reported in March that Gemini 2.5 Pro is the first release in the 2.5 series, which is claimed to be the most advanced reasoning AI model with "state-of-the-art" ranking across various benchmarks. Gemini 2.5 Pro got the top position on the large language model benchmarking tool LMArena, and received a score of 18.8% on Humanity's Last Exam, which has been specifically designed to test knowledge and reasoning capabilities. After being available only for Google AI Studio and Gemini app advanced users, the AI models have now become useful for developers. The tech major revealed that organisations like Snap Inc and SmartBear have already experimented with these models, in addition to developers like Spline and Rooms. In April, Google launched Gemini 2.5 Flash capable of real-time conversations, summaries, and extended interactions. It was available for both Google AI Studio and Vertex AI, paving the way for developers and companies to build AI agents. Following this, the company announced Gemini 2.5 Flash-Lite in preview. It claims that it is the most cost-efficient and fastest model in the series yet, which is efficient in coding, math, science, reasoning and multimodal benchmarks. It also supports a context length of up to one million tokens, which refer to words or pieces of a word used by AI systems.
[11]
Google has Supercharged Productivity Alongside Gemini: The Entry of Gemini 2.5 Marks the Beginning of a New Era of Efficiency.
At Google I/O 2025, Sundar Pichai unveiled Gemini 2.5 Pro, boasting ten times the performance of its predecessor. Google now processes 480 trillion tokens monthly, a fifty-fold increase from last year, with over 7 million developers utilizing Gemini. Integrated deeply into Android and Chrome OS, Gemini offers contextual assistance and automation.At the Google I/O 2025 keynote, CEO Sunder Pichai excitedly spoke of the rapid model progress of Gemini since their first-generation Gemini Pro model. Since then, Gemini 2.5 Pro today delivers 10 times the performance of the previous generation. What makes this stand out more is their infrastructure strength enables them to deliver dramatically faster models, even as the model prices climb down significantly. He further highlights the numbers wherein, this time last year, Google processed 9.7 trillion tokens a month across products and APIs, whilst today, they are processing 50 times more than that number, over 480 trillion. More than 7 million developers today use Gemini, which is five times more than what was garnered last year. The app Gemini now has over 400 million monthly active users, with a significant growth in engagement with the 2.5 series of models. These numbers are a clear indicator of Gemini 2.5's strength, deeming it as one of the most powerful multimodal models ever released by Google. This brand new version enhances performance across the board, possesses faster reasoning, deeper document understanding, and dramatically improved code generation. Google stands proud on its claims that Gemini 2.5 is capable of working through complex workflows that include summarizing multi-document reports and debugging intricate software projects with immaculacy. This model's rollout is closely associated with Google Workspace, bringing in AI-driven "Co-pilot for your digital life" rather than a mere AI-powered assistant, with its capabilities extending to Gmail, Docs, Sheets, and Slides. Via the "Help Me Write" and "Help Me Organize" features driven by Gemini 2.5, users can now curate detailed project briefs, rewrite proposals, extract spreadsheet insights, and much more on the list, making collaborations more efficient than ever. Gemini is no longer a mere Workspace upgrade; rather, it now exists in the very core of Android and Chrome OS. It can interpret on-screen content, provide contextual suggestions, and even automate device actions. Whether it's summarizing PDFs or drafting replies on chatting apps, Gemini is readily available to assist at the right moment, for its intelligence is now entangled in the depths of the Android experience. For businesses to use AI for internal processes, proprietary data, and domain-specific tasks, Google has also introduced Gemini for Workspace Enterprise and Gemini for Cloud AI. Organizations shall be able to incorporate Gemini into their workflow with optimum-level data privacy and governance controls; this move seemed like an insinuation of a challenge to Microsoft's Copilot in this new-age race of AI-enabled dominance. As AI grows and develops further into the global workplace, Google's bold push signifies not just a technological advancement but rather a strategic reshaping of how humans and machines shall join hands in the near future.
[12]
Google's Gemini 2.5 Stable Build Released : An AI That Can Do It All
What if the future of artificial intelligence wasn't just smarter -- but fundamentally more versatile? With the release of Gemini 2.5, Google has unveiled a new leap in AI technology, setting a new standard for what's possible. Imagine an AI capable of seamlessly analyzing text, audio, images, video, and even code -- all in a single workflow. This isn't just an incremental update; it's a bold redefinition of how AI can integrate into our lives, from transforming app development to decoding the complexities of multilingual communication. In an era where efficiency and adaptability are paramount, Gemini 2.5 doesn't just meet expectations -- it reshapes them. Matthew Berman explores how Gemini 2.5's multimodal capabilities and innovative training frameworks are pushing the boundaries of AI performance. From its ability to process sprawling datasets with a 1-million-token context window to its resource-efficient architecture, this release promises to empower developers, researchers, and businesses alike. But what truly sets Gemini 2.5 apart? Beyond its technical prowess, it's the model's real-world applications -- like analyzing intricate video content or assisting with complex coding tasks -- that make it a fantastic option. As we delve deeper, you'll discover why this release isn't just a milestone for Google but a pivotal moment for the AI landscape as a whole. Gemini 2.5 is engineered to process complex, multimodal inputs, including text, audio, images, video, and even code repositories. This versatility unlocks a wide array of applications, ranging from software development to video content analysis. Key features that distinguish Gemini 2.5 include: These features make Gemini 2.5 not only a high-performing model but also a resource-efficient solution, addressing the growing demand for scalable and versatile AI systems. Gemini 2.5 is built for speed, cost efficiency, and adaptability, making it suitable for a wide range of real-world applications. Its advanced capabilities excel in areas such as: For example, Gemini 2.5 can streamline app development workflows, generate coherent outputs for coding projects, or analyze complex video data to uncover patterns and trends. Its ability to handle long-context reasoning and multimodal interactions makes it an indispensable tool for developers, researchers, and businesses. Here is a selection of other guides from our extensive library of content you may find of interest on Google Gemini 2.5. The exceptional performance of Gemini 2.5 is rooted in its robust training framework. By using diverse datasets that include text, code, images, audio, and video, the models achieve a comprehensive understanding of various data modalities. Key training innovations include: These advancements enable Gemini 2.5 to deliver high-quality results while maintaining efficiency, making it a valuable asset for addressing complex AI challenges across industries. Google has prioritized safety and ethical considerations in the development of Gemini 2.5, implementing measures to ensure responsible AI usage. These include: These safeguards reflect Google's commitment to addressing concerns about data security, ethical AI use, and the potential risks associated with advanced AI technologies. The Gemini 2.5 models are powered by Google's TPU V5P architecture, which serves as the computational backbone for their advanced capabilities. This architecture enhances several critical aspects of the models, including: These technical advancements ensure that Gemini 2.5 remains a robust and reliable platform for tackling complex AI challenges with precision and speed. Despite its new features, Gemini 2.5 is not without limitations. Current challenges include: While these limitations highlight areas for improvement, ongoing research and development efforts are likely to address these challenges in future iterations. Even with these constraints, Gemini 2.5 remains a highly effective solution for most use cases. The practical applications of Gemini 2.5 underscore its versatility and adaptability. The models have been successfully employed in tasks such as: These examples highlight the diverse and demanding tasks that Gemini 2.5 can handle, making it a valuable resource for developers, researchers, and businesses seeking innovative AI solutions.
[13]
Gemini 2.5 Flash Hybrid Reasoning AI Optimized for AI Thinking for Efficiency
What if artificial intelligence could think only when you needed it to? Imagine a tool that seamlessly transitions between complex reasoning and straightforward processing, adapting to your specific needs without wasting resources. Enter Google's Gemini 2.5 Flash, a new AI model that redefines efficiency with its hybrid reasoning capabilities. By allowing developers to toggle between "thinking" and "non-thinking" modes, Gemini 2.5 Flash offers a level of control and adaptability that traditional AI systems simply can't match. Whether you're solving intricate problems or managing routine tasks, this innovation promises to deliver precision, scalability, and cost-efficiency -- all tailored to your workflow. In this coverage, Prompt Engineering explore how Gemini 2.5 Flash is reshaping the AI landscape with its thinking budget optimization, multimodal processing, and enhanced token capacities. You'll discover how its unique architecture eliminates the need for separate models, streamlining operations while reducing costs. But it's not without its limitations -- plateauing performance at higher token usage and capped reasoning budgets raise important questions about its scalability for resource-intensive projects. As we unpack its strengths and challenges, you'll gain a deeper understanding of whether Gemini 2.5 Flash is the right fit for your next AI endeavor. Sometimes, the real innovation lies in knowing when not to think. At the core of Gemini 2.5 Flash lies its hybrid reasoning model, a feature that distinguishes it from traditional AI systems. This capability enables you to toggle "thinking mode" on or off based on the complexity of the task. By managing the "thinking budget" -- the maximum number of tokens allocated for reasoning -- you can optimize the model's performance to suit specific use cases. This approach eliminates the need for separate models for reasoning-intensive and simpler tasks, streamlining workflows and reducing operational overhead. Whether you're addressing intricate problem-solving scenarios or routine data processing, the model's adaptability ensures optimal performance. The ability to fine-tune the reasoning process provides a significant advantage, allowing you to allocate resources efficiently while achieving high-quality results. Gemini 2.5 Flash is designed with cost-conscious developers in mind, offering a pricing structure that reflects its focus on affordability and performance. The model's pricing tiers are as follows: This competitive pricing positions Gemini 2.5 Flash as a cost-effective alternative to other leading AI models, such as OpenAI and DeepSync. By integrating proprietary hardware and software, Google ensures a strong performance-to-cost ratio, making the model an attractive option for projects that require scalability without sacrificing quality. This balance between affordability and capability makes it a practical choice for developers aiming to optimize their resources. Find more information on Hybrid Reasoning AI by browsing our extensive range of articles, guides and tutorials. In benchmark evaluations, Gemini 2.5 Flash ranks second overall on the Chatbot Arena leaderboard, trailing only OpenAI's O4 Mini in specific areas. However, it demonstrates significant improvements over its predecessor, Gemini 2.0 Flash, particularly in academic benchmarks. These advancements highlight the model's enhanced capabilities and its potential to deliver robust performance across various applications. While these results underscore its strengths, it is recommended that you test the model against your internal benchmarks to determine its suitability for your unique requirements. This hands-on evaluation will provide a clearer understanding of how Gemini 2.5 Flash can integrate into your workflows and meet your specific needs. One of the standout features of Gemini 2.5 Flash is its enhanced token capacity, which significantly expands its utility for developers. The model supports: These enhancements provide a substantial advantage for handling complex inputs and generating detailed outputs. Whether you're working on data-heavy projects or applications requiring extensive contextual understanding, Gemini 2.5 Flash offers the tools necessary to manage these challenges effectively. Gemini 2.5 Flash extends its capabilities to multimodal processing, supporting a variety of input types, including video, audio, and images. This versatility makes it a valuable tool for industries such as media analysis, technical documentation, and beyond. However, it is important to note that the model does not include image generation features, which may limit its appeal for creative applications. Despite this limitation, its ability to process diverse input types enhances its utility across a wide range of use cases. While Gemini 2.5 Flash excels in many areas, it is not without its limitations. These include: These constraints highlight areas where the model may fall short, particularly for developers requiring advanced reasoning capabilities or higher token limits. Understanding these limitations is crucial for making informed decisions about the model's applicability to your projects. Google's Gemini 2.5 Flash reflects a strategic focus on cost optimization, scalability, and accessibility, making advanced AI technology available to a broader audience. Its hybrid reasoning capabilities, enhanced token and context window capacities, and multimodal processing features position it as a versatile and scalable tool for developers. By balancing quality, cost, and latency, the model caters to a wide range of applications, from data analysis to technical problem-solving. For developers seeking practical solutions that combine flexibility, performance, and affordability, Gemini 2.5 Flash offers a compelling option. Its ability to adapt to diverse tasks and optimize resource allocation ensures that it can meet the demands of modern AI challenges effectively.
[14]
Google rolls out stable Gemini 2.5 Flash and Pro, previews 2.5 Flash-Lite model
On Tuesday, Google announced the expansion of its Gemini 2.5 AI lineup by officially releasing the 2.5 Pro and 2.5 Flash models as stable and widely accessible. The company also introduced a preview of Gemini 2.5 Flash-Lite, highlighting it as the quickest and most affordable model in the 2.5 series so far. Tulsee Doshi, Senior Director of Product Management for Gemini, credited user insights for refining the 2.5 Pro and Flash releases. Over recent weeks, organizations including Snap, SmartBear, Spline, and Rooms have already put these models into live use. Google rolled out the Flash-Lite preview, which surpasses its 2.0 predecessor in benchmarks covering programming, mathematical problem-solving, scientific analysis, logical reasoning, and handling multiple input formats. Designed to handle large volumes of data with minimal delay, Flash-Lite responds faster than the earlier 2.0 Flash and 2.0 Flash-Lite models, making it ideal for tasks like translation and text classification. It accommodates up to one million tokens, supports varied input types, hooks into tools like Google Search and code execution, and allows dynamic scaling of compute resources to match budget needs. You can access the Gemini 2.5 Flash-Lite preview and the stable 2.5 Pro and Flash models through Google AI Studio, Vertex AI, or the Gemini app, and Google has also embedded customized Flash and Flash-Lite versions within its Search platform.
Share
Copy Link
Google has announced the general availability of its Gemini 2.5 Pro and Flash AI models, along with the introduction of a new Gemini 2.5 Flash-Lite model, signaling a major push into the enterprise AI market.
Google has made a significant leap in the artificial intelligence arena by announcing the general availability of its Gemini 2.5 Pro and Flash AI models. This move marks a transition from experimental preview to production-ready status, signaling Google's confidence in the technology's ability to handle critical business applications 15. The tech giant is positioning itself to challenge OpenAI's dominance in the enterprise AI market.
Source: Ars Technica
The Gemini 2.5 family now includes three key models:
Gemini 2.5 Pro: Google's most capable model, excelling in complex reasoning, advanced code generation, and multimodal understanding. It can process up to one million tokens of context, equivalent to about 750,000 words 5.
Gemini 2.5 Flash: Designed for high-throughput enterprise tasks, balancing capability and efficiency 2.
Source: Geeky Gadgets
A distinguishing feature of Google's approach is the emphasis on "reasoning" or "thinking" capabilities. This architecture allows models to process problems more deliberately before responding, giving developers unprecedented control over AI behavior through adjustable "thinking budgets" 5.
Google has restructured its pricing strategy to compete across market segments:
Several major companies are already leveraging Gemini 2.5 models in production:
Google's move comes at a strategic time, weeks after OpenAI faced scrutiny over the safety and reliability of its latest models. This creates an opportunity for Google to position itself as a more stable, enterprise-focused alternative 5. The company is clearly aiming to strengthen its position in the AI arms race, offering a comprehensive suite of AI tools that span from premium reasoning capabilities to budget-conscious automation options.
The launch of these production-ready models signifies Google's commitment to AI development and its ambition to capture a larger share of the enterprise AI market. As businesses increasingly demand scalable and reliable AI systems, Google's Gemini 2.5 family is poised to play a significant role in shaping the future of enterprise AI applications 5.
With these advancements, Google is not only challenging OpenAI's market leadership but also setting new standards for AI model capabilities, efficiency, and pricing in the rapidly evolving landscape of artificial intelligence.
NASA and IBM have developed Surya, an open-source AI model that can predict solar flares and space weather, potentially improving the protection of Earth's critical infrastructure from solar storms.
5 Sources
Technology
1 hr ago
5 Sources
Technology
1 hr ago
Meta introduces an AI-driven voice translation feature for Facebook and Instagram creators, enabling automatic dubbing of content from English to Spanish and vice versa, with plans for future language expansions.
8 Sources
Technology
17 hrs ago
8 Sources
Technology
17 hrs ago
OpenAI CEO Sam Altman reveals plans for GPT-6, focusing on memory capabilities to create more personalized and adaptive AI interactions. The upcoming model aims to remember user preferences and conversations, potentially transforming the relationship between humans and AI.
2 Sources
Technology
17 hrs ago
2 Sources
Technology
17 hrs ago
Chinese AI companies DeepSeek and Baidu are making waves in the global AI landscape with their open-source models, challenging the dominance of Western tech giants and potentially reshaping the AI industry.
2 Sources
Technology
1 hr ago
2 Sources
Technology
1 hr ago
A comprehensive look at the emerging phenomenon of 'AI psychosis', its impact on mental health, and the growing concerns among experts and tech leaders about the psychological risks associated with AI chatbots.
3 Sources
Technology
1 hr ago
3 Sources
Technology
1 hr ago