2 Sources
[1]
Writer releases Palmyra X5, delivers near GPT-4 performance at 75% lower cost
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Writer, the enterprise generative AI company valued at $1.9 billion, today released Palmyra X5, a new large language model (LLM) featuring an expansive 1-million-token context window that promises to accelerate the adoption of autonomous AI agents in corporate environments. The San Francisco-based company, which counts Accenture, Marriott, Uber, and Vanguard among its hundreds of enterprise customers, has positioned the model as a cost-efficient alternative to offerings from industry giants like OpenAI and Anthropic, with pricing set at $0.60 per million input tokens and $6 per million output tokens. "This model really unlocks the agentic world," said Matan-Paul Shetrit, Director of Product at Writer, in an interview with VentureBeat. "It's faster and cheaper than equivalent large context window models out there like GPT-4.1, and when you combine it with the large context window and the model's ability to do tool or function calling, it allows you to start really doing things like multi-step agentic flows." AI economics breakthrough: How Writer trained a powerhouse model for just $1 million Unlike many competitors, Writer trained Palmyra X5 with synthetic data for approximately $1 million in GPU costs -- a fraction of what other leading models require. This cost efficiency represents a significant departure from the prevailing industry approach of spending tens or hundreds of millions on model development. "Our belief is that tokens in general are becoming cheaper and cheaper, and the compute is becoming cheaper and cheaper," Shetrit explained. "We're here to solve real problems, rather than nickel and diming our customers on the pricing." The company's cost advantage stems from proprietary techniques developed over several years. In 2023, Writer published research on "becoming self-instruct," which introduced early stopping criteria for minimal instruct tuning. According to Shetrit, this allows Writer to "cut costs significantly" during the training process. "Unlike other foundational shops, our view is that we need to be effective. We need to be efficient here," Shetrit said. "We need to provide the fastest, cheapest models to our customers, because ROI really matters in these cases." Million-token marvel: The technical architecture powering Palmyra X5's speed and accuracy Palmyra X5 can process a full million-token prompt in approximately 22 seconds and execute multi-turn function calls in around 300 milliseconds -- performance metrics that Writer claims enable "agent behaviors that were previously cost- or time-prohibitive." The model's architecture incorporates two key technical innovations: a hybrid attention mechanism and a mixture of experts approach. "The hybrid attention mechanism...introduces attention mechanism that inside the model allows it to focus on the relevant parts of the inputs when generating each output," Shetrit said. This approach accelerates response generation while maintaining accuracy across the extensive context window. On benchmark tests, Palmyra X5 achieved notable results relative to its cost. On OpenAI's MRCR 8-needle test -- which challenges models to find eight identical requests hidden in a massive conversation -- Palmyra X5 scored 19.1%, compared to 20.25% for GPT-4.1 and 17.63% for GPT-4o. It also places eighth in coding on the BigCodeBench benchmark with a score of 48.7. These benchmarks demonstrate that while Palmyra X5 may not lead every performance category, it delivers near-flagship capabilities at significantly lower costs -- a trade-off that Writer believes will resonate with enterprise customers focused on ROI. From chatbots to business automation: How AI agents are transforming enterprise workflows The release of Palmyra X5 comes shortly after Writer unveiled AI HQ earlier this month -- a centralized platform for enterprises to build, deploy, and supervise AI agents. This dual product strategy positions Writer to capitalize on growing enterprise demand for AI that can execute complex business processes autonomously. "In the age of agents, models offering less than 1 million tokens of context will quickly become irrelevant for business-critical use cases," said Writer CTO and co-founder Waseem AlShikh in a statement. Shetrit elaborated on this point: "For a long time, there's been a large gap between the promise of AI agents and what they could actually deliver. But at Writer, we're now seeing real-world agent implementations with major enterprise customers. And when I say real customers, it's not like a travel agent use case. I'm talking about Global 2000 companies, solving the gnarliest problems in their business." Early adopters are deploying Palmyra X5 for various enterprise workflows, including financial reporting, RFP responses, support documentation, and customer feedback analysis. One particularly compelling use case involves multi-step agentic workflows, where an AI agent can flag outdated content, generate suggested revisions, share them for human approval, and automatically push approved updates to a content management system. This shift from simple text generation to process automation represents a fundamental evolution in how enterprises deploy AI -- moving from augmenting human work to automating entire business functions. Cloud expansion strategy: AWS partnership brings Writer's AI to millions of enterprise developers Alongside the model release, Writer announced that both Palmyra X5 and its predecessor, Palmyra X4, are now available in Amazon Bedrock, Amazon Web Services' fully managed service for accessing foundation models. AWS becomes the first cloud provider to deliver fully managed models from Writer, significantly expanding the company's potential reach. "Seamless access to Writer's Palmyra X5 will enable developers and enterprises to build and scale AI agents and transform how they reason over vast amounts of enterprise data -- leveraging the security, scalability, and performance of AWS," said Atul Deo, Director of Amazon Bedrock at AWS, in the announcement. The AWS integration addresses a critical barrier to enterprise AI adoption: the technical complexity of deploying and managing models at scale. By making Palmyra X5 available through Bedrock's simplified API, Writer can potentially reach millions of developers who lack the specialized expertise to work with foundation models directly. Self-learning AI: Writer's vision for models that improve without human intervention Writer has staked a bold claim regarding context windows, announcing that 1 million tokens will be the minimum size for all future models it releases. This commitment reflects the company's view that large context is essential for enterprise-grade AI agents that interact with multiple systems and data sources. Looking ahead, Shetrit identified self-evolving models as the next major advancement in enterprise AI. "The reality is today, agents do not perform at the level we want and need them to perform," he said. "What I think is realistic is as users come to AI HQ, they start doing this process mapping...and then you layer on top of that, or within it, the self-evolving models that learn from how you do things in your company." These self-evolving capabilities would fundamentally change how AI systems improve over time. Rather than requiring periodic retraining or fine-tuning by AI specialists, the models would learn continuously from their interactions, gradually improving their performance for specific enterprise use cases. "This idea that one agent can rule them all is not realistic," Shetrit noted when discussing the varied needs of different business teams. "Even two different product teams, they have so many such different ways of doing work, the PMs themselves." Enterprise AI's new math: How Writer's $1.9B strategy challenges OpenAI and Anthropic Writer's approach contrasts sharply with that of OpenAI and Anthropic, which have raised billions in funding but focus more on general-purpose AI development. Writer has instead concentrated on building enterprise-specific models with cost profiles that enable widespread deployment. This strategy has attracted significant investor interest, with the company raising $200 million in Series C funding last November at a $1.9 billion valuation. The round was co-led by Premji Invest, Radical Ventures, and ICONIQ Growth, with participation from strategic investors including Salesforce Ventures, Adobe Ventures, and IBM Ventures. According to Forbes, Writer has a remarkable 160% net retention rate, indicating that customers typically expand their contracts by 60% after initial adoption. The company reportedly has over $50 million in signed contracts and projects this will double to $100 million this year. For enterprises evaluating generative AI investments, Writer's Palmyra X5 presents a compelling value proposition: powerful capabilities at a fraction of the cost of competing solutions. As the AI agent ecosystem matures, the company's bet on cost-efficient, enterprise-focused models could position it advantageously against better-funded competitors that may not be as attuned to business ROI requirements. "Our goal is to drive widespread agent adoption across our customer base as quickly as possible," Shetrit emphasized. "The economics are straightforward -- if we price our solution too high, enterprises will simply compare the cost of an AI agent versus a human worker and may not see sufficient value. To accelerate adoption, we need to deliver both superior speed and significantly lower costs. That's the only way to achieve large-scale deployment of these agents within major enterprises." In an industry often captivated by technical capabilities and theoretical performance ceilings, Writer's pragmatic focus on cost efficiency might ultimately prove more revolutionary than another decimal point of benchmark improvement. As enterprises grow increasingly sophisticated in measuring AI's business impact, the question may shift from "How powerful is your model?" to "How affordable is your intelligence?" -- and Writer is betting its future that economics, not just capabilities, will determine AI's enterprise winners.
[2]
Writer announces Palmyra X5 LLM with 1M-token context window to power AI agents - SiliconANGLE
Writer announces Palmyra X5 LLM with 1M-token context window to power AI agents Generative artificial intelligence startup Writer Inc. today released its newest state-of-the-art enterprise-focused large language model Palmyra X5, an adaptive reasoning model that features a 1 million-token context window. The company also said that it teamed up with Amazon Web Services Inc. to announce the availability of the new model on Amazon Bedrock, AWS's fully managed service for building AI applications by providing access to LLMs and machine learning models. In an interview with SiliconANGLE, Matan-Paul Shetrit, director of product management at Writer, said Palmyra X5 was trained entirely using synthetic data and required just $1 million in graphics processing unit hours. "Writer has been a leader from very early on in the use of synthetic data, as well as techniques that allow us to cut down costs, like stopping training if we don't see returns, which lets us push model pricing way down," Shetrit said. With its 1 million-token large-context window, the model can ingest and comprehend extensive codebases or large numbers of documents simultaneously. It also enables multistep reasoning across actions, making it ideal for powering agentic AI workflows. "Palmyra X5 is the newest model that comes with a 1 million-token context window, which is one of the largest commercially available, and it's amongst the fastest and cheapest large-context LLMs," Shetrit said. "This is critical when it comes to building multi-agentic systems, where the responses need to be continuously fed back into the context window across multistep workflows." Shetrit said that going forward, Writer will incorporate the 1 million-token context window into all of Writer's Palmyra family of models. The model can read an entire 1 million-token prompt in about 22 seconds, and it can return individual function-calling turns in about 300 milliseconds. X5 performs with a pricing of 60 cents per 1 million input tokens and $6 per 1 million output tokens. Between the low latency and low cost, it permits the model to provide high-speed performance across long contexts while allowing enterprise users to perform numerous actions without costing too much. This is essential, especially for agentic workflows, which can quickly become costly and slow with multistep and multiturn AI generations. "If you're thinking about enterprise use cases, they're extremely complex and interacting with multiple third-party, external or internal systems, often in multistep flows," Shetrit said. "Our focus isn't on solving for one- or two-agent workflows. It's on enabling enterprises to manage 10, 50, or even 100,000 agents all working alongside employees to drive real transformational change." The release of Palmyra X5 follows Writer's launch of AI HQ, a new centralized hub for teams to easily build, deploy and orchestrate AI agents. Shetrit described Palmyra X5 and the company's family of other Palmyra models as powerful "brains" for agents. He said X5 delivers on the company's vision for helping companies adapt to the agentic AI trend. "Palmyra represents a critical step forward for companies seeking to embed generative AI into their enterprise workflows," said David Cushman, executive research leader at HFS Research Ltd. "Palmyra is purpose-built for agent development, API function calling, multimodal processing and custom enterprise AI applications. More importantly, it's integrated into Writer's end-to-end platform that balances scalability with transparency -- a core demand from regulators and stakeholders alike." Hundreds of enterprise companies already use Writer's Palmyra family of LLMs internally and externally to power their AI capabilities, including Accenture Plc, Marriott International Inc., Prudential Financial Inc., Qualcomm Inc., Uber Technologies Inc., Salesforce Inc. and The Vanguard Group Inc.
Share
Copy Link
Writer, an enterprise AI company, releases Palmyra X5, a large language model with a 1-million-token context window. The model promises near GPT-4 performance at 75% lower cost, potentially accelerating the adoption of AI agents in corporate environments.
Writer, a San Francisco-based enterprise generative AI company valued at $1.5 billion, has unveiled Palmyra X5, a new large language model (LLM) that promises to revolutionize the AI landscape for businesses 1. This innovative model features a massive 1-million-token context window, positioning it as a powerful tool for accelerating the adoption of autonomous AI agents in corporate environments 12.
Palmyra X5 is being marketed as a cost-effective alternative to offerings from industry leaders like OpenAI and Anthropic. Writer has set competitive pricing at $0.60 per million input tokens and $6 per million output tokens 1. This pricing strategy, combined with the model's impressive capabilities, aims to make advanced AI more accessible to a broader range of enterprises.
The model incorporates two key technical innovations: a hybrid attention mechanism and a mixture of experts approach 1. These advancements allow Palmyra X5 to process a full million-token prompt in approximately 22 seconds and execute multi-turn function calls in around 300 milliseconds 12. Such performance metrics enable "agent behaviors that were previously cost- or time-prohibitive" 1.
In benchmark tests, Palmyra X5 has shown impressive results. On OpenAI's MRCR 8-needle test, it scored 19.5 compared to 20.0 for GPT-4 and 17.5 for GPT-4-0314 1. Additionally, it ranks eighth in coding on the BigCodeBench benchmark with a score of 48.5 1.
Unlike many competitors, Writer trained Palmyra X5 using synthetic data for approximately $1 million in GPU costs 12. This represents a significant departure from the industry norm of spending tens or hundreds of millions on model development. The company's cost advantage stems from proprietary techniques developed over several years, including "becoming self-instruct" research published in 2023 1.
The release of Palmyra X5 comes shortly after Writer unveiled AI HQ, a centralized platform for enterprises to build, deploy, and supervise AI agents 1. This dual product strategy positions Writer to capitalize on growing enterprise demand for AI that can execute complex business processes autonomously.
Early adopters are deploying Palmyra X5 for various enterprise workflows, including:
One particularly compelling use case involves multi-step agentic workflows, where an AI agent can flag outdated content, generate suggested revisions, share them for human approval, and automatically push approved updates to a content management system 1.
Alongside the model release, Writer announced that both Palmyra X5 and its predecessor, Palmyra X4, are now available in Amazon Bedrock, Amazon Web Services' fully managed service for accessing foundation models 12. This partnership with AWS significantly expands Writer's potential reach, making its models accessible to millions of enterprise developers 1.
The introduction of Palmyra X5 represents a significant step forward in the enterprise AI landscape. As Waseem AlShikh, Writer's CTO and co-founder, stated, "In the age of agents, models offering less than 1 million tokens of context will quickly become irrelevant for business-critical use cases" 1. This shift from simple text generation to process automation marks a fundamental evolution in how enterprises deploy AI, moving from augmenting human work to automating entire business functions 12.
With hundreds of enterprise companies already using Writer's Palmyra family of LLMs, including Accenture, Marriott, Uber, and Vanguard, the release of Palmyra X5 is poised to further accelerate the adoption of advanced AI solutions in the corporate world 12.
NVIDIA announces significant upgrades to its GeForce NOW cloud gaming service, including RTX 5080-class performance, improved streaming quality, and an expanded game library, set to launch in September 2025.
9 Sources
Technology
6 hrs ago
9 Sources
Technology
6 hrs ago
As nations compete for dominance in space, the risk of satellite hijacking and space-based weapons escalates, transforming outer space into a potential battlefield with far-reaching consequences for global security and economy.
7 Sources
Technology
22 hrs ago
7 Sources
Technology
22 hrs ago
OpenAI updates GPT-5 to make it more approachable following user feedback, sparking debate about AI personality and user preferences.
6 Sources
Technology
14 hrs ago
6 Sources
Technology
14 hrs ago
A pro-Russian propaganda group, Storm-1679, is using AI-generated content and impersonating legitimate news outlets to spread disinformation, raising concerns about the growing threat of AI-powered fake news.
2 Sources
Technology
22 hrs ago
2 Sources
Technology
22 hrs ago
A study reveals patients' increasing reliance on AI for medical advice, often trusting it over doctors. This trend is reshaping doctor-patient dynamics and raising concerns about AI's limitations in healthcare.
3 Sources
Health
14 hrs ago
3 Sources
Health
14 hrs ago