Curated by THEOUTPOST
On Sat, 14 Dec, 8:01 AM UTC
2 Sources
[1]
Cohere's smallest, fastest R-series model excels at RAG, reasoning in 23 languages
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Proving its intention to support a wide range of enterprise use cases -- including those that don't require expensive, resource-intensive large language models (LLMs) -- AI startup Cohere has released Command R7B, the smallest and fastest in its R model series. Command R7B is built to support fast prototyping and iteration and uses retrieval-augmented generation (RAG) to improve its accuracy. The model features a context length of 128K and supports 23 languages. It outperforms others in its class of open-weights models -- Google's Gemma, Meta's Llama, Mistral's Ministral -- in tasks including math and coding, Cohere says. "The model is designed for developers and businesses that need to optimize for the speed, cost-performance and compute resources of their use cases," Cohere co-founder and CEO Aidan Gomez writes in a blog post announcing the new model. Outperforming competitors in math, coding, RAG Cohere has been strategically focused on enterprises and their unique use cases. The company introduced Command-R in March and the powerful Command R+ in April, and has made upgrades throughout the year to support speed and efficiency. It teased Command R7B as the "final" model in its R series, and says it will release model weights to the AI research community. Cohere noted that a critical area of focus when developing Command R7B was to improve performance on math, reasoning, code and translation. The company appears to have succeeded in those areas, with the new smaller model topping the HuggingFace Open LLM Leaderboard against similarly-sized open-weight models including Gemma 2 9B, Ministral 8B and Llama 3.1 8B. Further, the smallest model in the R series outperforms competing models in areas including AI agents, tool use and RAG, which helps improve accuracy by grounding model outputs in external data. Cohere says Command R7B excels at conversational tasks including tech workplace and enterprise risk management (ERM) assistance; technical facts; media workplace and customer service support; HR FAQs; and summarization. Cohere also notes that the model is "exceptionally good" at retrieving and manipulating numerical information in financial settings. All told, Command R7B ranked first, on average, in important benchmarks including instruction-following evaluation (IFeval); big bench hard (BBH); graduate-level Google-proof Q&A (GPQA); multi-step soft reasoning (MuSR); and massive multitask language understanding (MMLU). Removing unnecessary call functions Command R7B can use tools including search engines, APIs and vector databases to expand its functionality. Cohere reports that the model's tool use performs strongly against competitors in the Berkeley Function-Calling Leaderboard, which evaluates a model's accuracy in function calling (connecting to external data and systems). Gomez points out that this proves its effectiveness in "real-world, diverse and dynamic environments" and removes the need for unnecessary call functions. This can make it a good choice for building "fast and capable" AI agents. For instance, Cohere points out, when functioning as an internet-augmented search agent, Command R7B can break complex questions down into subgoals, while also performing well with advanced reasoning and information retrieval. Because it is small, Command R7B can be deployed on lower-end and consumer CPUs, GPUs and MacBooks, allowing for on-device inference. The model is available now on the Cohere platform and HuggingFace. Pricing is $0.0375 per 1 million input tokens and $0.15 per 1 million output tokens. "It is an ideal choice for enterprises looking for a cost-efficient model grounded in their internal documents and data," writes Gomez.
[2]
Cohere Launches Command R7B to Disrupt Enterprise AI Market
The model is suitable for deployment on low-end GPUs, CPUs, and even MacBooks. Canadian AI startup, Cohere has launched Command R7B, the smallest model in its R series of large language models (LLMs), targeting businesses with a focus on speed, cost efficiency, and flexibility. The model is suitable for deployment on low-end GPUs, CPUs, and even MacBooks. It supports a context length of 128k and offers features such as retrieval-augmented generation (RAG) with native inline citations, multilingual capabilities, and performance across math, code, and reasoning tasks. Cohere highlighted its suitability for enterprise use cases such as customer service and HR. "Command R7B balances efficiency with performance, allowing businesses to deploy high-quality AI solutions on affordable infrastructure," Cohere said in its announcement. The model has demonstrated strong results on the HuggingFace Open LLM Leaderboard and outperforms competitors in tasks related to RAG, tool use, and AI agents. Its performance has been evaluated against multiple benchmarks, including ChatRAGBench, StrategyQA, and the Berkeley Function-Calling Leaderboard. Command R7B is accessible via the Cohere Platform and HuggingFace, with the model's weights released for use by the AI research community. Cohere is offering the model at $0.0375 per million input tokens. Command 7B joins other small language models released this week, including Microsoft's Phi-4 and Google's PaliGemma 2. Cohere recently launched Rerank 3.5 to enhance search relevance and content ranking for enterprises, offering multilingual support in over 100 languages, including Arabic, Chinese, English, French, German, Hindi, Japanese, Korean, Portuguese, Russian, and Spanish. The company recently secured a $240 million investment from the Canadian government to build a multibillion-dollar AI data centre in Canada. Founded in 2019, Cohere specialises in developing LLMs for business applications. Unlike some of its counterparts, such as OpenAI and Google, Cohere focuses on enterprise solely rather than pursuing artificial general intelligence (AGI). Earlier this year, Cohere reached a valuation of $5.5 billion raising $500 million in their Series D funding round.
Share
Share
Copy Link
Cohere introduces Command R7B, the smallest model in its R series, designed for enterprise use with a focus on efficiency, performance, and versatility across multiple languages and tasks.
Cohere, the Canadian AI startup, has launched Command R7B, the smallest and fastest model in its R series of large language models (LLMs). This release marks a significant step in Cohere's strategy to cater to a wide range of enterprise use cases, particularly those that don't require resource-intensive models 1.
Command R7B boasts impressive specifications:
The model excels in various tasks, including math, coding, reasoning, and translation. It has demonstrated strong performance in AI agents, tool use, and RAG applications 1.
Command R7B has shown remarkable results on several benchmarks:
Cohere emphasizes Command R7B's suitability for various enterprise use cases:
Command R7B can integrate with tools such as search engines, APIs, and vector databases. It performs strongly on the Berkeley Function-Calling Leaderboard, demonstrating its effectiveness in real-world, diverse environments 1.
The model is now available on the Cohere platform and HuggingFace. Pricing is set at $0.0375 per million input tokens and $0.15 per million output tokens 1. Cohere has also released the model weights to the AI research community 2.
Cohere's launch of Command R7B aligns with its focus on enterprise AI solutions. The company, founded in 2019, recently secured a $240 million investment from the Canadian government for a multibillion-dollar AI data center. With a valuation of $5.5 billion after its Series D funding round, Cohere is positioning itself as a key player in the enterprise AI market 2.
As the AI landscape continues to evolve, Cohere's Command R7B represents a significant advancement in balancing efficiency and performance for businesses seeking to deploy high-quality AI solutions on affordable infrastructure.
Reference
[2]
Cohere's research arm releases Aya Expanse, a family of multilingual AI models that outperform leading open-source alternatives, aiming to bridge the global language divide in AI technology.
3 Sources
3 Sources
Canadian AI startup Cohere announces a strategic shift towards developing tailored AI models for enterprise users, moving away from the race to build larger foundation models.
3 Sources
3 Sources
Cohere, a prominent AI startup, recently secured $500 million in funding, reaching a $5.5 billion valuation. However, the company subsequently laid off 20% of its workforce, signaling a strategic realignment in the competitive AI landscape.
13 Sources
13 Sources
Cohere, a Canada-based AI startup, introduces North, a secure AI platform designed to enhance enterprise workflows with LLM-powered tools, advanced search capabilities, and customizable AI agents, prioritizing data security and privacy for regulated industries.
4 Sources
4 Sources
Cohere launches Embed 3, an advanced multimodal AI model that integrates text and image embeddings, setting new standards for enterprise search and multilingual retrieval tasks.
2 Sources
2 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved