Curated by THEOUTPOST
On Thu, 13 Feb, 12:03 AM UTC
2 Sources
[1]
Is Perplexity's Sonar really more 'factual' than its AI rivals? See for yourself
The company claims its newly upgraded model is number one in user satisfaction and speed - but its methodology is unclear. AI search engine Perplexity says its latest release goes above and beyond for user satisfaction -- especially compared to OpenAI's GPT-4o. On Tuesday, Perplexity announced a new version of Sonar, its proprietary model. Based on Meta's open-source Llama 3.3 70B, the updated Sonar "is optimized for answer quality and user experience," the company says, having been trained to improve the readability and accuracy of its answers in search mode. Also: The billion-dollar AI company no one is talking about - and why you should care Perplexity claims Sonar scored higher than GPT-4o mini and Claude models on factuality and readability. The company defines factuality as a measure of "how well a model can answer questions using facts that are grounded in search results, and its ability to resolve conflicting or missing information." However, there isn't an external benchmark to measure this. Instead, Perplexity displays several screenshot examples of side-by-side answers from Sonar and competitor models including GPT-4o and Claude 3.5 Sonnet. They do, in my opinion, differ in directness, completion, and scannability, often favoring Sonar's cleaner formatting (a subjective preference) and higher number of citations -- though that doesn't speak directly to source quality, only quantity. The sources a chatbot cites are also influenced by the publisher and media partner agreements of its parent company, which Perplexity and OpenAI each have. More importantly, the examples don't include the queries themselves, only the answers, and Perplexity does not clarify a methodology on how it provoked or measured the responses -- differences between queries, the number of queries run, etc. -- instead leaving the comparisons up to individuals to "see the difference." ZDNET has reached out to Perplexity for comment. Perplexity says that online A/B testing revealed that users were much more satisfied and engaged with Sonar than with GPT-4o mini, Claude 3.5 Haiku, and Claude 3.5 Sonnet, but it didn't expand on the specifics of these results. Also: The work tasks people use Claude AI for most, according to Anthropic "Sonar significantly outperforms models in its class, like GPT-4o mini and Claude 3.5 Haiku, while closely matching or exceeding the performance of frontier models like GPT-4o and Claude 3.5 Sonnet for user satisfaction," Perplexity's announcement states. According to Perplexity, Sonar's speed of 1,200 tokens per second enables it to answer queries almost instantly and work 10 times faster than Gemini 2.0 Flash. Testing showed Sonar surpassing GPT-4o mini and Claude 3.5 Haiku "by a substantial margin," but the company doesn't clarify the details of that testing. The company also says Sonar outperforms more expensive frontier models like Claude 3.5 Sonnet "while closely approaching the performance of GPT-4o." Sonar did beat its two competitors, among others, on academic benchmark tests IFEval and MMLU, which evaluate how well a model follows user instructions and its grasp of "world knowledge" across disciplines. Also: Cerebras CEO on DeepSeek: Every time computing gets cheaper, the market gets bigger Want to try it for yourself? The upgraded Sonar is available for all Pro users, who can make it their default model in their settings or access it through the Sonar API.
[2]
Perplexity Launches Sonar for Pro Users; Performance on Par with GPT-4o, Claude 3.5 Sonnet
Sonar is powered by Cerebras Inference, which claims to be the world's fastest AI inference engine. Perplexity, an AI search engine startup, announced that its in-house model, Sonar, will be available to all Pro users on the platform. Now, users with the Perplexity Pro plan can make Sonar the default model via settings. Sonar is built on top of Meta's open-source Llama 3.3 70B. It is powered by Cerebras Inference, which claims to be the world's fastest AI inference engine. The model is capable of producing 1200 tokens per second. "We optimised Sonar across two critical dimensions that strongly correlate with user satisfaction - answer factuality and readability," Perplexity announced, indicating that Sonar significantly improves the base Llama model on these aspects. Perplexity revealed that their evaluations found that Sonar outperforms OpenAI's GPT-4o mini and Anthropic's Claude 3.5 Haiku and offers performance parity with the bigger models GPT-4o and Claude 3.5 Sonnet. Furthermore, Perplexity said Sonar is 10 times faster than Google's Gemini 2.0 Flash. Recently, French AI startup Mistral revealed its app, Le Chat, which claimed to be the fastest AI assistant in the competition. During our testing, we found it to be faster than all other models. Gemini 2.0 Flash, on the other hand, came in second. Like Perplexity's Sonar, Mistral's Le Chat is also powered by Cerebras Inference. Recently, Perplexity also announced the availability of the powerful DeepSeek-R1 model on the platform, hosted on servers in the United States. A few weeks ago, Perplexity announced that the Sonar API is available in two variants: the Sonar and the Sonar Pro. The company also called it the most affordable API in the market. The company said Sonar Pro is "ideal for multi-step tasks requiring deep understanding and context retention". Moreover, it provides "in-depth answers" with twice the citations of Sonar. The Pro version costs $3 per million input tokens, $15 per million output tokens, and $5 per 1,000 searches, with multiple searches allowed. The Sonar plan is simpler. It charges $1 per million tokens for input and output and $5 per 1,000 searches, with only one search per request.
Share
Share
Copy Link
Perplexity launches Sonar, an AI model built on Meta's Llama 3.3 70B, claiming superior performance and user satisfaction compared to competitors like GPT-4 and Claude 3.5. The company's methodology and comparisons, however, raise questions about transparency and objectivity.
Perplexity, an AI search engine startup, has launched Sonar, its proprietary AI model, for all Pro users on its platform 1. Built on Meta's open-source Llama 3.3 70B and powered by Cerebras Inference, Sonar claims to outperform leading AI models in factuality, readability, and speed 2.
Perplexity asserts that Sonar surpasses OpenAI's GPT-4o mini and Anthropic's Claude 3.5 Haiku in performance, while matching or exceeding GPT-4o and Claude 3.5 Sonnet in user satisfaction 1. The company also states that Sonar operates at 1,200 tokens per second, making it nearly 10 times faster than Google's Gemini 2.0 Flash 2.
According to Perplexity, Sonar outperformed its competitors in academic benchmark tests such as IFEval and MMLU, which evaluate instruction-following capabilities and general knowledge across disciplines 1. The company claims that A/B testing revealed higher user satisfaction and engagement with Sonar compared to rival models 1.
While Perplexity provides screenshot examples comparing Sonar's outputs to those of competitor models, the methodology behind these comparisons remains unclear 1. The company does not disclose details about the queries used, the number of tests conducted, or the specific metrics for measuring factuality and readability 1.
Perplexity has made the Sonar API available in two variants: Sonar and Sonar Pro 2. The company touts it as the most affordable API in the market, with Sonar Pro costing $3 per million input tokens, $15 per million output tokens, and $5 per 1,000 searches 2. The standard Sonar plan charges $1 per million tokens for both input and output, with a $5 per 1,000 searches fee 2.
Sonar's launch comes amid fierce competition in the AI model space. French startup Mistral recently introduced Le Chat, which also uses Cerebras Inference and claims to be the fastest AI assistant available 2. Perplexity has also added the DeepSeek-R1 model to its platform, hosted on U.S. servers, further diversifying its AI offerings 2.
The introduction of Sonar and its claimed performance metrics could potentially shake up the AI model landscape. However, the lack of standardized, independent benchmarks for factuality and user satisfaction in AI search engines makes it challenging to verify these claims objectively 1. As the AI industry continues to evolve rapidly, the need for transparent and standardized evaluation methods becomes increasingly apparent.
Reference
[2]
Analytics India Magazine
|Perplexity Launches Sonar for Pro Users; Performance on Par with GPT-4o, Claude 3.5 SonnetPerplexity introduces Sonar API, offering developers and enterprises access to real-time AI search capabilities. The launch includes two tiers - Sonar and Sonar Pro - with competitive pricing and advanced features, challenging major players in the AI industry.
5 Sources
5 Sources
Perplexity AI introduces a free Deep Research feature, powered by DeepSeek R1, offering comprehensive research capabilities to all users and potentially disrupting the AI-powered search market.
12 Sources
12 Sources
AI search platforms Perplexity and You.com have integrated DeepSeek's R1 model, allowing users to experience the advanced AI without the security risks associated with China-based servers.
2 Sources
2 Sources
A comprehensive overview of the latest AI models from xAI, Anthropic, OpenAI, and Google, highlighting their unique features, capabilities, and accessibility.
2 Sources
2 Sources
Google has launched Gemini 2.5 Pro, its latest AI model boasting advanced reasoning capabilities, multimodality, and improved performance across various benchmarks. This release marks a significant step in the ongoing AI race among tech giants.
39 Sources
39 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved