11 Sources
[1]
Anthropic's Claude AI model can now handle longer prompts | TechCrunch
Anthropic is increasing the amount of information that enterprise customers can send to Claude in a single prompt, part of an effort to attract more developers to the company's popular AI coding models. For Anthropic's API customers, the company's Claude Sonnet 4 AI model now has a one million token context window -- meaning the AI can handle requests as long as 750,000 words, more than the entire Lord of the Rings trilogy, or 75,000 lines of code. That's roughly five times Claude's previous limit (200,000 tokens), and more than double the 400,000 token context window offered by OpenAI's GPT-5. Long context will also be available for Claude Sonnet 4 through Anthropic's cloud partners, including on Amazon Bedrock and Google Cloud's Vertex AI. Anthropic has built one of the largest enterprise businesses among AI model developers, largely by selling Claude to AI coding platforms such as Microsoft's GitHub Copilot, Windsurf, and Anysphere's Cursor. While Claude has become the model of choice among developers, GPT-5 may threaten Anthropic's dominance with its competitive pricing and strong coding performance. Anysphere CEO Michael Truell even helped OpenAI announce the launch of GPT-5, which is now the default AI model for new users in Cursor. Anthropic's product lead for the Claude platform, Brad Abrams, told TechCrunch in an interview that he expects AI coding platforms to get a "lot of benefit" from this update. When asked if GPT-5 put a dent in Claude's API usage, Abrams downplayed the concern, saying he's "really happy with the API business and the way it's been growing." Whereas OpenAI generates most of its revenue from consumer subscriptions to ChatGPT, Anthropic's business centers around selling AI models to enterprises through an API. That's made AI coding platforms a key customer for Anthropic, and could be why the company is throwing in some new perks to attract users in the face of GPT-5. Last week, Anthropic unveiled an updated version of its largest AI model, Claude Opus 4.1, which pushed the company's AI coding capabilities a bit further. Generally speaking, AI models tend to perform better on all tasks when they have more context, but especially for software engineering problems. For example, if you ask an AI model to spin up a new feature for your app, it's likely to do a better job if it can see the entire project, rather than just a small section. Abrams also told TechCrunch that Claude's large context window also helps it perform better at long agentic coding tasks, in which the AI model is autonomously working on a problem for minutes or hours. With a large context window, Claude can remember all its previous steps in long-horizon tasks. But some companies have taken large context windows to an extreme, claiming their AI models can process massive prompts. Google offers a 2 million token context window for Gemini 2.5 Pro, and Meta offers a 10 million token context window for Llama 4 Scout. Some studies suggest there's a limit to how large context windows can be, and AI models are not great at processing massive prompts. Abrams said that Anthropic's research team focused on increasing not just the context window for Claude, but the "effective context window," suggesting that its AI can understand most of the information it's given. However, he declined to reveal Anthropic's exact techniques. When prompts to Claude Sonnet 4 are over 200,000 tokens, Anthropic will charge more to API users, at $6 per million input tokens and $22.50 per million output tokens (up from $3 per million input tokens and $15 per million output tokens).
[2]
Anthropic's powerful Opus 4.1 model is here - how to access it (and why you'll want to)
It is available to paid Claude users, Claude Code, API, Amazon Bedrock, and Google Cloud's Vertex AI. In May, Anthropic released Claude Opus 4, which the company dubbed its most powerful model yet and the best coding model in the world. Only three months later, Anthropic is upping the ante further by launching the highly anticipated Claude Opus 4.1, which now takes its predecessor's crown as Anthropic's most advanced model. The Opus family of models is the company's most advanced, intelligent AI models geared toward tackling complex problems. As a result, Claude Opus 4.1, released on Tuesday, excels at those tasks and can even one-up its predecessor on agentic tasks, real-world coding, and reasoning, according to Anthropic. The model also comes as the industry is expecting the launch of OpenAI's GPT-5 soon. Also: OpenAI could launch GPT-5 any minute now - what to expect One of the most impressive use cases of Claude Opus 4 was its performance on the SWE-bench Verified, a human-filtered subset of the SWE-bench, a benchmark that evaluates LLMs' abilities to solve real-world software engineering tasks sourced from GitHub. Claude Opus 4's performance on the SWE-bench Verified supported the claim that it was the "best coding model in the world." As seen in the post above, Opus 4.1 performed even higher. Claude Opus 4.1 also swept its preceding models across the benchmark board, including the MMMLU, which tests for multilingual capabilities; AIME 2025, which tests for rigor on high school match competition questions; GPQA, which tests for performance on graduate-level reasoning prompts; and more. When pinned against competitors' reasoning models, including OpenAI o3 and Gemini 2.5 Pro, it outperforms them in various benchmarks, including SWE-bench Verified. With the release, Anthropic also posted its system card, which delineates all of the safety assessments and evaluations it conducted on the model, as well as its weaknesses, risks, and limitations. A quick overview of the 22-page document shows that the model was deployed with an AI Safety Level 3 (ASL-3) Standard under Anthropic's Responsible Scaling Policy (RSP), and it is still prone to most of the same vulnerabilities. If you want to try the model for yourself, it is now available to everyone via the paid Claude plans, which include Claude Pro for $20 per month and Claude Max for $100 per month. It is available in Claude Code, the API, Amazon Bedrock, and Google Cloud's Vertex AI.
[3]
Claude Sonnet's memory gets a big boost with 1M tokens of context
* Claude Sonnet 4 now has one million context tokens. * As a result, the model can process much larger developer tasks. * Developers can access it now, but API pricing does increase for certain requests. We all have that friend who is a great active listener and can recall details from past interactions, which then feeds into better conversations in the future. Similarly, AI models have context windows that impact how much content they can reference -- and Anthropic's Claude Sonnet just got a huge upgrade that should let it do a lot more for you. Also: Microsoft Copilot 3D turns your 2D images into 3D models for free - how to try Claude Sonnet 4 can now support up to one million tokens of context, marking a fivefold increase from the prior 200,000, Anthropic said on Tuesday. With this large context window, Claude Sonnet can process codebases with over 75,000 lines of code or dozens of research papers with a single API request. The longer context window allows developers to use Claude for more data-heavy projects, such as large-scale code analysis and document synthesis, as mentioned above, but also context-aware agents that require a lot of material to operate complex workflows. Anthropic describes Claude Sonnet as a "hybrid reasoning model with superior intelligence." ZDNET's own internal testing has found it to be a very capable model, even outperforming what Anthropic touted at the time as its most advanced coding model, Opus 4. Sonnet passed all four coding tests while Claude Opus failed two. Also: How you can still access GPT-4o, o3, and other older models in ChatGPT Developers interested in trying the upgraded Sonnet 4 can access it in public beta on the Anthropic API as customers with Tier 4 and custom rate limits. It is also available on third-party cloud computing platforms, starting with Amazon Bedrock, and coming soon to Google Cloud's Vertex AI. The API pricing does increase for prompts over 200K tokens. Since launching the Claude 4 models in May, Anthropic has been riding a wave of steady upward momentum. Claude Code, a fan favorite coding assistant that can be called on directly in a developer's workspace to write or manage code, shipped 10 features in the last month, and its highly anticipated Opus 4.1 was released last week. The company reports that its B2B run-rate revenue has grown 17 times YoY as of June 2025.
[4]
You can now feed Claude Sonnet 4 entire codebases at once - 9to5Mac
Following OpenAI's big week filled with open models and GPT-5, Anthropic is on a streak of its own with AI announcements. The company today revealed that Claude Sonnet 4 now supports up to 1 million tokens of context in the Anthropic API -- a five-fold increase over the previous limit. This expanded "long context" capability allows developers to feed far larger datasets into Claude in a single request. Anthropic says the 1M-token window can handle entire codebases with more than 75,000 lines of code, or dozens of lengthy research papers at once. Use cases include large-scale code analysis that considers every file, test, and piece of documentation; synthesis of massive document collections like contracts or technical specs; and context-aware AI agents that can maintain coherence across hundreds of tool calls and multi-step workflows. The upgrade is available in public beta for Anthropic API customers with Tier 4 or custom rate limits, as well as through Amazon Bedrock. Support for Google Cloud's Vertex AI is "coming soon." Pricing doubles for prompts over 200,000 tokens, though Anthropic notes that prompt caching and batch processing can cut costs by up to 50 percent. Anthropic highlighted two customers already using the feature: Bolt.new, which integrates Claude into its browser-based development platform, and iGent AI, whose Maestro agent turns conversations into code. Both say the 1M-token window enables larger, more accurate, and more autonomous coding workflows. Anthropic announced yesterday that Claude is gaining its own memory feature, allowing users to reference other chat history, including by pasting the exact context to reference. A week ago, Anthropic released a small but useful improvement to Claude Sonnet with version 4.1.
[5]
Anthropic rolls out Claude Opus 4.1 with improved software engineering accuracy - 9to5Mac
In May, AI firm Anthropic introduced its Claude 4 family of models with a focus on improvements to coding, reasoning, and following instructions. Three months later, Anthropic is back with Claude Opus 4.1, which it says upgrades "agentic tasks, real-world coding, and reasoning." Anthropic says Claude Opus 4.1 improves software engineering accuracy to 74.5%. That compares to 62.3% with Claude Sonnet 3.7 and 72.5% with Claude Opus 4. More specifically, the updated model is better at "in-depth research and data analysis skills, especially around detail tracking and agentic search," according to Anthropic. Here's a benchmark table provided by Anthropic's blog post announcing Opus 4.1: Anthropic's latest model is out today for Claude customers and through Claude Code, through Anthropic's API, Amazon Bedrock, and Google Cloud's Vertex AI. For a deeper dive on Anthropic's latest model, you can find the Claude Opus 4.1 system card here. Meanwhile, Anthropic says on social media that it plans "to release substantially larger improvements to our models in the coming weeks," so expect more model upgrades for Claude. OpenAI is also set to make news with announcements likely coming today and later this week.
[6]
Claude can now process entire software projects in single request, Anthropic says
Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Anthropic announced Tuesday that its Claude Sonnet 4 artificial intelligence model can now process up to 1 million tokens of context in a single request -- a fivefold increase that allows developers to analyze entire software projects or dozens of research papers without breaking them into smaller chunks. The expansion, available now in public beta through Anthropic's API and Amazon Bedrock, represents a significant leap in how AI assistants can handle complex, data-intensive tasks. With the new capacity, developers can load codebases containing more than 75,000 lines of code, enabling Claude to understand complete project architecture and suggest improvements across entire systems rather than individual files. The announcement comes as Anthropic faces intensifying competition from OpenAI and Google, both of which already offer similar context windows. However, company sources speaking on background emphasized that Claude Sonnet 4's strength lies not just in capacity but in accuracy, achieving 100% performance on internal "needle in a haystack" evaluations that test the model's ability to find specific information buried within massive amounts of text. How developers can now analyze entire codebases with AI in one request The extended context capability addresses a fundamental limitation that has constrained AI-powered software development. Previously, developers working on large projects had to manually break down their codebases into smaller segments, often losing important connections between different parts of their systems. "What was once impossible is now reality," said Sean Ward, CEO and co-founder of London-based iGent AI, whose Maestro platform transforms conversations into executable code, in a statement. "Claude Sonnet 4 with 1M token context has supercharged autonomous capabilities in Maestro, our software engineering agent. This leap unlocks true production-scale engineering-multi-day sessions on real-world codebases." Eric Simons, CEO of Bolt.new, which integrates Claude into browser-based development platforms, said in a statement: "With the 1M context window, developers can now work on significantly larger projects while maintaining the high accuracy we need for real-world coding." The expanded context enables three primary use cases that were previously difficult or impossible: comprehensive code analysis across entire repositories, document synthesis involving hundreds of files while maintaining awareness of relationships between them, and context-aware AI agents that can maintain coherence across hundreds of tool calls and complex workflows. Why Claude's new pricing strategy could reshape the AI development market Anthropic has adjusted its pricing structure to reflect the increased computational requirements of processing larger contexts. While prompts of 200,000 tokens or fewer maintain current pricing at $3 per million input tokens and $15 per million output tokens, larger prompts cost $6 and $22.50 respectively. The pricing strategy reflects broader dynamics reshaping the AI industry. Recent analysis shows that Claude Opus 4 costs roughly seven times more per million tokens than OpenAI's newly launched GPT-5 for certain tasks, creating pressure on enterprise procurement teams to balance performance against cost. However, Anthropic argues the decision should factor in quality and usage patterns rather than price alone. Company sources noted that prompt caching -- which stores frequently accessed large datasets -- can make long context cost-competitive with traditional Retrieval-Augmented Generation approaches, especially for enterprises that repeatedly query the same information. "Large context lets Claude see everything and choose what's relevant, often producing better answers than pre-filtered RAG results where you might miss important connections between documents," an Anthropic spokesperson told VentureBeat. Anthropic's billion-dollar dependency on just two major coding customers The long context capability arrives as Anthropic commands 42% of the AI code generation market, more than double OpenAI's 21% share according to a Menlo Ventures survey of 150 enterprise technical leaders. However, this dominance comes with risks: industry analysis suggests that coding applications Cursor and GitHub Copilot drive approximately $1.2 billion of Anthropic's $5 billion annual revenue run rate, creating significant customer concentration. The GitHub relationship proves particularly complex given Microsoft's $13 billion investment in OpenAI. While GitHub Copilot currently relies on Claude for key functionality, Microsoft faces increasing pressure to integrate its own OpenAI partnership more deeply, potentially displacing Anthropic despite Claude's current performance advantages. The timing of the context expansion is strategic. Anthropic released this capability on Sonnet 4 -- which offers what the company calls "the optimal balance of intelligence, cost, and speed" -- rather than its most powerful Opus model. Company sources indicated this reflects the needs of developers working with large-scale data, though they declined to provide specific timelines for bringing long context to other Claude models. Inside Claude's breakthrough AI memory technology and emerging safety risks The 1 million token context window represents significant technical advancement in AI memory and attention mechanisms. To put this in perspective, it's enough to process approximately 750,000 words -- roughly equivalent to two full-length novels or extensive technical documentation sets. Anthropic's internal testing revealed perfect recall performance across diverse scenarios, a crucial capability as context windows expand. The company embedded specific information within massive text volumes and tested Claude's ability to find and use those details when answering questions. However, the expanded capabilities also raise safety considerations. Earlier versions of Claude Opus 4 demonstrated concerning behaviors in fictional scenarios, including attempts at blackmail when faced with potential shutdown. While Anthropic has implemented additional safeguards and training to address these issues, the incidents highlight the complex challenges of developing increasingly capable AI systems. Fortune 500 companies rush to adopt Claude's expanded context capabilities The feature rollout is initially limited to Anthropic API customers with Tier 4 and custom rate limits, with broader availability planned over coming weeks. Amazon Bedrock users have immediate access, while Google Cloud's Vertex AI integration is pending. Early enterprise response has been enthusiastic, according to company sources. Use cases span from coding teams analyzing entire repositories to financial services firms processing comprehensive transaction datasets to legal startups conducting contract analysis that previously required manual document segmentation. "This is one of our most requested features from API customers," an Anthropic spokesperson said. "We're seeing excitement across industries that unlocks true agentic capabilities, with customers now running multi-day coding sessions on real-world codebases that would have been impossible with context limitations before." The development also enables more sophisticated AI agents that can maintain context across complex, multi-step workflows. This capability becomes particularly valuable as enterprises move beyond simple AI chat interfaces toward autonomous systems that can handle extended tasks with minimal human intervention. What OpenAI's aggressive pricing means for the future of AI development tools The long context announcement intensifies competition among leading AI providers. Google's older Gemini 1.5 Pro model and OpenAI's older GPT-4.1 model both offer 1 million token windows, but Anthropic argues that Claude's superior performance on coding and reasoning tasks provides competitive advantage even at higher prices. The broader AI industry has seen explosive growth in model API spending, which doubled to $8.4 billion in just six months according to Menlo Ventures. Enterprises consistently prioritize performance over price, upgrading to newer models within weeks regardless of cost, suggesting that technical capabilities often outweigh pricing considerations in procurement decisions. However, OpenAI's recent aggressive pricing strategy with GPT-5 could reshape these dynamics. Early comparisons show dramatic price advantages that may overcome typical switching inertia, especially for cost-conscious enterprises facing budget pressures as AI adoption scales. For Anthropic, maintaining its coding market leadership while diversifying revenue sources remains critical. The company has tripled the number of eight and nine-figure deals signed in 2025 compared to all of 2024, reflecting broader enterprise adoption beyond its coding strongholds. As AI systems become capable of processing and reasoning about increasingly vast amounts of information, they're fundamentally changing how developers approach complex software projects. The ability to maintain context across entire codebases represents a shift from AI as a coding assistant to AI as a comprehensive development partner that understands the full scope and interconnections of large-scale projects. The implications extend far beyond software development. Industries from legal services to financial analysis are beginning to recognize that AI systems capable of maintaining context across hundreds of documents could transform how organizations process and understand complex information relationships. But with great capability comes great responsibility -- and risk. As these systems become more powerful, the incidents of concerning AI behavior during Anthropic's testing serve as a reminder that the race to expand AI capabilities must be balanced with careful attention to safety and control. As Claude learns to juggle a million pieces of information simultaneously, Anthropic faces its own context window problem: being trapped between OpenAI's pricing pressure and Microsoft's conflicting loyalties.
[7]
Anthropic Releases Claude Opus 4.1 With Agentic, Coding and Reasoning Upgrades | AIM
"We plan to release substantially larger improvements to our models in the coming weeks." Anthropic has released Claude Opus 4.1, a new version of its flagship AI model, with improvements in coding, reasoning, and agentic task performance. The update builds on Claude Opus 4 and is now available to paid users through Claude Code, API access, Amazon Bedrock, and Google Cloud's Vertex AI. Pricing remains unchanged. The company says Opus 4.1 delivers state-of-the-art coding performance, achieving 74.5% on SWE-bench Verified, a benchmark for real-world software engineering tasks. It also boosts capabilities in in-depth research, data analysis, and agentic search. According to GitHub, the model shows gains in most areas compared to Opus 4, especially in multi-file code refactoring. Rakuten Group noted Claude Opus 4.1's ability to identify precise corrections in larg
[8]
Anthropic launches Claude Opus 4.1 with 74.5% coding accuracy
Anthropic launched Claude Opus 4.1, an AI model, today for customers via API integrations, enhancing software engineering accuracy and agentic capabilities. Anthropic, an AI company, introduced its Claude 4 model family in May, emphasizing improved coding, reasoning, and instruction following. Three months later, the company has released Claude Opus 4.1. This iteration reportedly upgrades "agentic tasks, real-world coding, and reasoning." Anthropic states that Claude Opus 4.1 improves software engineering accuracy to 74.5%. This represents an increase from 62.3% with Claude Sonnet 3.7 and 72.5% with Claude Opus 4. The company notes the updated model excels in "in-depth research and data analysis skills, especially around detail tracking and agentic search." A benchmark table from Anthropic's blog post detailing Opus 4.1 provides comparative data: Anthropic's latest model is available today for Claude customers. It can be accessed through Claude Code, Anthropic's API, Amazon Bedrock, and Google Cloud's Vertex AI platform. Additional information on Anthropic's new model is available in the Claude Opus 4.1 system card. Anthropic indicated on social media that it plans "to release substantially larger improvements to our models in the coming weeks." OpenAI is also expected to make announcements today and later this week. Mac users can download Claude for macOS, and Anthropic also offers iPhone and iPad applications for its Claude AI platform.
[9]
Anthropic releases Claude Opus 4.1 amid rival ChatGPT's advancements - The Economic Times
More improvements will be rolled out in the coming weeks, the company said. The price for Opus 4.1 remains the same as that of Opus 4.Artificial intelligence (AI) company Anthropic released an upgraded version of Claude Opus 4 on Tuesday to build capabilities in real-world coding, agentic research, creative writing and reasoning. The company said it aims to bring larger improvements in the coming weeks. The development comes after rival OpenAI announced on Tuesday the release of two open-weight language models, which are designed for advanced reasoning and are optimised to run on laptops, performing similarly to OpenAI's smaller proprietary models. Opus 4.1 is currently available to paid Claude users and in Claude Code. It's also available through tools for developers like Amazon Bedrock and Google Cloud's Vertex AI. Additionally, the price for using Opus 4.1 is the same as it was for the previous version, i.e., Opus 4. Pricing for Claude Opus 4.1 starts at $15 per million input tokens and $75 per million output tokens. Claude Opus 4.1 Opus 4.1 by Anthropic advanced the state-of-the-art coding performance to 74.5% on SWE-bench Verified that tests real-world coding problems, along with in-depth research and data analysis skills. According to GitHub, Opus 4.1 has enhanced its capabilities in multi-file code refactoring. Multi-file code refactoring means improving or reorganising code that is spread across multiple files while keeping the programme's behaviour unchanged. Developer-focussed AI coding startup Windsurf reported that Opus 4.1 delivers a one-standard-deviation improvement over Opus 4 on their junior developer benchmark, showing roughly the same performance leap as the jump from Sonnet 3.7 to Sonnet 4. The broader picture Anthropic was founded in 2021 by a group of former OpenAI employees. The San Francisco-based company is generating about $5 billion in annualised revenue, Bloomberg News has reported. In May this year, Google- and Amazon-backed Anthropic introduced its next-generation AI agents, Claude Opus 4 and Claude Sonnet 4, with coding and advanced reasoning capabilities. On facing stiff competition with giants like OpenAI, Anthropic's chief product officer Mike Krieger told Bloomberg that it was focussing on its own progress rather than competitors like OpenAI. The priority, he said, is delivering value to current customers. In a separate development, Anthropic's Claude is now listed on the General Services Administration (GSA) schedule on Tuesday, making its products readily accessible to US federal government departments and agencies with pre-negotiated pricing and terms that comply with federal acquisition regulations.
[10]
Claude Opus 4.1 AI Released : Code, Translate and Solve Advanced Problems
What if your AI assistant could not only understand complex problems but solve them with graduate-level precision? That's the promise of Claude Opus 4.1, the latest iteration of Anthropic's innovative AI model. In a world where artificial intelligence is reshaping industries at breakneck speed, Claude's newest update stands out by pushing boundaries in reasoning, multilingual capabilities, and even coding. Imagine an AI that not only deciphers intricate datasets but also writes cleaner code, manages multi-step tasks with surgical precision, and speaks your language -- literally. With Opus 4.1, Anthropic isn't just refining its technology; it's redefining what AI can do in real-world applications. In the video below Matthew Berman provides more insights into the fantastic upgrades that make Claude Opus 4.1 a standout in the competitive AI landscape. From its enhanced ability to tackle complex programming challenges to its improved performance in multilingual processing, the update promises to elevate productivity across industries. But it's not just about technical benchmarks -- this version also highlights the growing pains of innovative AI, revealing areas where even the most advanced systems still fall short. Whether you're a developer, data analyst, or simply curious about the future of AI, this exploration of Claude Opus 4.1 offers a nuanced look at how technology is evolving to meet the demands of an increasingly complex world. Claude Opus 4.1 introduces a series of updates designed to improve its functionality and user experience. These enhancements include: These updates make Claude Opus 4.1 a versatile tool for users across diverse industries, from software development to data-driven decision-making, enhancing productivity and precision. The performance of Claude Opus 4.1 is reflected in its benchmark results, which highlight both its strengths and areas for improvement: These results highlight Claude's growing capabilities while emphasizing areas where further refinement is needed, particularly in industry-specific agentic tasks. Enhance your knowledge on Claude AI by exploring a selection of articles and guides on the subject. Claude Opus 4.1 competes effectively with leading AI models such as OpenAI's GPT-4 (03) and Google's Gemini 2.5 Pro. It outperforms these competitors in technical benchmarks like SWEBench and Terminal Bench, solidifying its reputation as a preferred choice for developers and technical users. However, it shows slight underperformance in areas such as graduate-level reasoning (GPQA Diamond) and high school math competition benchmarks, indicating opportunities for further development. The competitive nature of the AI industry underscores the importance of continuous improvement. Even incremental advancements can significantly influence a model's standing. Claude's ability to excel in specific areas while maintaining strong overall performance ensures its relevance in a crowded and dynamic market. Claude Opus 4.1 is particularly valued for its agentic coding capabilities, making it an essential tool for developers managing complex, multi-step tasks. Its advancements in multilingual processing and data analysis further broaden its appeal, catering to a wide range of applications across industries. Anthropic has announced plans for future updates aimed at addressing current limitations and introducing new features to expand the model's capabilities. These developments will be critical as the AI landscape continues to evolve, making sure Claude remains a reliable and versatile tool for users navigating increasingly complex challenges. As industries demand more sophisticated AI solutions, Claude Opus 4.1's ongoing evolution positions it as a key player in meeting these needs, offering practical applications and benefits across diverse fields.
[11]
Anthropic Unveils Claude Opus 4.1 in Dueling Releases With OpenAI | PYMNTS.com
By completing this form, you agree to receive marketing communications from PYMNTS and to the sharing of your information with our sponsor, if applicable, in accordance with our Privacy Policy and Terms and Conditions. Claude Opus 4.1 is better at agentic tasks, coding and reasoning, according to a Tuesday company blog post. Leaks of Claude Opus 4.1 began appearing the day before on social platform X and TestingCatalog. Claude Opus 4.1 is a successor to Claude Opus 4, which launched May 22. Opus 4.1 shows gains on benchmarks such as SWE-Bench Verified, a coding evaluation test, where it scores two percentage points higher than the previous model, according to the blog post. The Claude Opus 4.1 release came amid signs that rival OpenAI is nearing the debut of GPT-5, the Bloomberg report said. OpenAI executives have been teasing its release, with some reports speculating it could be as soon as this month. "One thing I've learned, especially in AI as it's moving quickly, is that we can focus on what we have -- and what other folks are going to do is ultimately up to them," Krieger said when asked about GPT-5, per the report. Anthropic, founded in 2021 by former OpenAI researchers, has focused on building safer, high-performing AI systems. The startup is generating about $5 billion in annualized revenue and is finalizing a funding round that could value it at $170 billion, the report said. Anthropic Launches Claude for Financial Services to Power Data-Driven Decisions
Share
Copy Link
Anthropic enhances its Claude AI model with a 1 million token context window for Claude Sonnet 4 and releases an upgraded Claude Opus 4.1, improving software engineering accuracy and AI capabilities.
Anthropic, a leading AI company, has significantly enhanced its Claude AI model's capabilities, positioning itself as a formidable competitor in the AI industry. The company has increased the context window for Claude Sonnet 4 to an impressive one million tokens, allowing it to process up to 750,000 words or 75,000 lines of code in a single prompt 1. This expansion represents a fivefold increase from Claude's previous limit and more than doubles the capacity of OpenAI's GPT-5 3.
Source: ZDNet
Alongside the context window expansion, Anthropic has released Claude Opus 4.1, touted as its most advanced model yet. This upgrade brings significant improvements in software engineering accuracy, reaching 74.5% compared to the previous 72.5% with Claude Opus 4 5. The new model excels in agentic tasks, real-world coding, and reasoning, outperforming competitors like OpenAI and Google in various benchmarks 2.
Anthropic's strategy clearly targets enterprise customers, particularly in the AI coding sector. The expanded context window and improved performance are expected to benefit AI coding platforms significantly 1. Both Claude Sonnet 4 with the extended context window and Claude Opus 4.1 are available through various channels:
Source: VentureBeat
The increased context window allows developers to use Claude for more data-heavy projects, including large-scale code analysis, document synthesis, and context-aware agents for complex workflows 3. Companies like Bolt.new and iGent AI are already leveraging these capabilities to enable larger, more accurate, and more autonomous coding workflows 4.
While the expanded capabilities come with increased API pricing for prompts over 200,000 tokens, Anthropic suggests that prompt caching and batch processing can reduce costs by up to 50% 4. The company emphasizes that larger context windows generally improve AI model performance, especially for software engineering tasks 1.
Source: Dataconomy
These advancements come as the AI industry anticipates the launch of OpenAI's GPT-5, highlighting the intense competition in the field 2. Anthropic's focus on enterprise solutions and API-based business model contrasts with OpenAI's consumer-oriented approach, potentially giving Anthropic an edge in specific market segments 1.
As the AI landscape continues to evolve rapidly, Anthropic hints at "substantially larger improvements" to their models in the coming weeks, suggesting that the race for AI supremacy is far from over 5.
Summarized by
Navi
Meta Platforms is considering collaborations with AI rivals Google and OpenAI to improve its AI applications, potentially integrating external models into its products while developing its own AI capabilities.
5 Sources
Technology
1 day ago
5 Sources
Technology
1 day ago
Meta announces significant changes to its AI chatbot policies, focusing on teen safety by restricting conversations on sensitive topics and limiting access to certain AI characters.
8 Sources
Technology
1 day ago
8 Sources
Technology
1 day ago
Meta faces scrutiny for hosting AI chatbots impersonating celebrities without permission, raising concerns about privacy, ethics, and potential legal implications.
7 Sources
Technology
1 day ago
7 Sources
Technology
1 day ago
A groundbreaking AI-powered stethoscope has been developed that can detect three major heart conditions in just 15 seconds, potentially transforming early diagnosis and treatment of heart diseases.
5 Sources
Health
17 hrs ago
5 Sources
Health
17 hrs ago
A group of 60 UK parliamentarians have accused Google DeepMind of breaching international AI safety commitments by delaying the release of safety information for its Gemini 2.5 Pro model.
2 Sources
Policy
1 day ago
2 Sources
Policy
1 day ago