Curated by THEOUTPOST
On Wed, 14 May, 12:02 AM UTC
2 Sources
[1]
Guardian agents: New approach could reduce AI hallucinations to below 1%
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Hallucination is a risk that limits the real-world deployment of enterprise AI. Many organizations have attempted to solve the challenge of hallucination reduction with various approaches, each with varying degrees of success. Among the many vendors that have been working for the last several years to reduce the risk is Vectara. The company got its start as an early pioneer in grounded retrieval, which is better known today by the acronym Retrieval Augmented Generation (RAG). An early promise of RAG was that it could help reduce hallucinations by sourcing information from provided content. While RAG is helpful as a hallucination reduction approach, hallucinations still occur even with RAG. Among existing industry solutions most solutions focus on detecting hallucinations or implementing preventative guardrails, Vectara has unveiled a fundamentally different approach: automatically identifying, explaining and correcting AI hallucinations through what it calls guardian agents inside of a new service called the Vectara Hallucination Corrector. The guardian agents are functionally software components that monitor and take protective actions within AI workflows. Instead of just applying rules inside of an LLM, the promise of guardian agents is to apply corrective measures in an agentic AI approach that improves workflows. Vectara's approach makes surgical corrections while preserving the overall content and providing detailed explanations of what was changed and why. The approach appears to deliver meaningful results. According to Vectara, the system can reduce hallucination rates for smaller language models under 7 billion parameters, to less than 1%. "As enterprises are implementing more agentic workflows, we all know that hallucinations are still an issue with LLMs and how that is going to exponentially amplify the negative impact of making mistakes in an agentic workflow is kind of scary for enterprises," Eva Nahari, chief product officer at Vectara told VentureBeat in an exclusive interview. "So what we have set out as a continuation of our mission to build out trusted AI and enable the full potential of gen AI for enterprise... is this new track of releasing guardian agents." Every enterprise wants to have accurate AI, that's not a surprise. It's also no surprise that there are many different options for reducing hallucinations. RAG approaches help to reduce hallucinations by providing grounded responses from content but can still yield inaccurate results. One of the more interesting implementations of RAG is one from the Mayo Clinic which uses a 'reverse RAG' approach to limit hallucinations. Improving data quality as well as how vector data embeddings are created is another approach to improving accuracy. Among the many vendors working on that approach is database vendor MongoDB which recently acquired advanced embedding and retrieval model vendor Voyage AI. Guardrails, which are available from many vendors including Nvidia and AWS among others, help to detect risky outputs and can help with accuracy in some cases. IBM actually has a set of its Granite open-source models known as Granite Guardian that directly integrate guardrails as a series of fine-tuning instructions, to reduce risky outputs. Using reasoning to validate output is another potential solution. AWS claims that its Bedrock Automated Reasoning approach catches 100% of hallucinations, though that claim is difficult to validate. Startup Oumi offers another approach, validating claims made by AI on a sentence by sentence basis by validating source materials with an open-source technology called HallOumi. While there is merit to all the other approaches to hallucination reduction, Vectara claims its approach is different. Rather than just identifying if a hallucination is present and then either flagging or rejecting the content, the guardian agent approach actually corrects the issue. Nahari emphasized that the guardian agent takes action. "It's not just a learning on something," she said. "It's taking an action on behalf of someone, and that makes it an agent." The guardian agent is a multi-stage pipeline rather than a single model. Suleman Kazi, machine learning tech lead at Vectara told VentureBeat that the system comprises three key components: a generative model, a hallucination detection model and a hallucination correction model. This agentic workflow allows for dynamic guardrailing of AI applications, addressing a critical concern for enterprises hesitant to fully embrace generative AI technologies. Rather than wholesale elimination of potentially problematic outputs, the system can make minimal, precise adjustments to specific terms or phrases. Here's how it works: The nuanced correction capabilities are critically important. Understanding the context of the query and source materials can make the difference between an answer being accurate or being a hallucination. When discussing the nuances of hallucination correction, Kazi provided a specific example to illustrate why blanket hallucination correction isn't always appropriate. He described a scenario where an AI is processing a science fiction book that describes the sky as red, instead of the typical blue. In this context, a rigid hallucination correction system might automatically "correct" the red sky to blue, which would be incorrect for the creative context of a science fiction narrative. The example was used to demonstrate that hallucination correction needs contextual understanding. Not every deviation from expected information is a true hallucination - some are intentional creative choices or domain-specific descriptions. This highlights the complexity of developing an AI system that can distinguish between genuine errors and purposeful variations in language and description. Alongside its guardian agent, Vectara is releasing HCMBench, an open-source evaluation toolkit for hallucination correction models. This benchmark provides standardized ways to evaluate how well different approaches correct hallucinations. The goal of the benchmark is to help the community at large, as well as to help enable enterprises to evaluate hallucination correction claims accuracy, including those from Vectara. The toolkit supports multiple metrics including HHEM, Minicheck, AXCEL and FACTSJudge, providing comprehensive evaluation of hallucination correction effectiveness. "If the community at large wants to develop their own correction models, they can use that benchmark as an evaluation data set to improve their models," Kazi said. For enterprises navigating the risks of AI hallucinations, Vectara's approach represents a significant shift in strategy. Instead of just implementing detection systems or abandoning AI in high-risk use cases, companies can now consider a middle path: implementing correction capabilities. The guardian agent approach also aligns with the trend toward more complex, multi-step AI workflows. Enterprises looking to implement these approaches should consider: With hallucination correction technologies maturing, enterprises may soon be able to deploy AI in previously restricted use cases while maintaining the accuracy standards required for critical business operations.
[2]
Vectara launches Hallucination Corrector to increase the reliability of enterprise AI - SiliconANGLE
Vectara launches Hallucination Corrector to increase the reliability of enterprise AI Artificial intelligence agent and assistant platform provider Vectara Inc. today announced the launch of a new Hallucination Corrector directly integrated into its service, designed to detect and mitigate costly, unreliable responses from enterprise AI models. Hallucinations, when generative AI large language models confidently provide false information, have long plagued the industry. For traditional models, it's estimated that they occur at a rate of about 3% to 10% of queries on average, depending on the model. The recent advent of reasoning AI models, which break down complex questions into step-by-step solutions to "think" through them, has led to a noted increase in the hallucination rate. According to a report from Vectara, DeepSeek-R1, a reasoning model, hallucinates significantly more at 14.3% than its predecessor DeepSeek R3 at 3.9%. Similarly, OpenAI's GPT-o1, also a reasoning model, jumped to a 2.4% rate from GPT-4o at 1.5%. New Scientist published a similar report and found even higher rates with the same and other reasoning models. "While LLMs have recently made significant progress in addressing the issue of hallucinations, they still fall distressingly short of the standards for accuracy that are required in highly regulated industries like financial services, healthcare, law and many others," said Vectara founder and Chief Executive Amr Awadallah. In its initial testing, Vectara said the Hallucination Corrector reduced hallucination rates in enterprise AI systems to about 0.9%. It works in conjunction with the company's already widely used Hughes Hallucination Evaluation Model, which provides a way to compare responses to source documents and identify if statements are accurate at runtime. The HHEM scores the answer against the source with a probability score between 1 and 0, where 0 means completely inaccurate - a total hallucination - and 1 for perfect accuracy. For example, a 0.98 means that there's a 98% chance the answer is likely highly accurate. HHEM is available on Hugging Face and received over 250,000 downloads last month, making it one of the most popular hallucination detectors on the platform. In the case of a factually inconsistent response, the Corrector provides a detailed output including an explanation of why the statement is a hallucination and a corrected version incorporating minimal changes for accuracy. The company said it automatically uses the corrected output in summaries for end-users, but experts can use the full explanation and suggested fixes for testing applications to refine or fine-tune their models and guardrails to combat hallucinations. It can also show the original summary, but use corrections info to flag potential uses while offering the corrected summary as an optional fix. In the case of LLM answers that fall into the category of misleading but not quite outright false, the Hallucination Corrector can work to refine the response to reduce its uncertainty core according to the customer's settings.
Share
Share
Copy Link
Vectara introduces a novel approach to reduce AI hallucinations below 1% using guardian agents, potentially transforming enterprise AI adoption by automatically identifying, explaining, and correcting inaccuracies.
Vectara, a pioneer in AI technology, has unveiled a groundbreaking solution to address one of the most significant challenges in enterprise AI adoption: hallucinations. The company's new Hallucination Corrector, powered by "guardian agents," promises to reduce hallucination rates to below 1% for smaller language models under 7 billion parameters 1.
AI hallucinations occur when large language models confidently provide false information. Traditional models typically experience hallucination rates between 3% to 10%, while newer reasoning AI models have shown even higher rates. For instance, DeepSeek-R1, a reasoning model, has been reported to hallucinate at a rate of 14.3% 2.
Vectara's Hallucination Corrector employs a multi-stage pipeline comprising three key components:
This agentic workflow allows for dynamic guardrailing of AI applications, addressing a critical concern for enterprises hesitant to fully embrace generative AI technologies 1.
Unlike other solutions that focus on detecting hallucinations or implementing preventative guardrails, Vectara's approach takes corrective action. The system makes minimal, precise adjustments to specific terms or phrases, preserving the overall content while providing detailed explanations of what was changed and why 1.
Vectara emphasizes the importance of contextual understanding in hallucination correction. Not every deviation from expected information is a true hallucination; some may be intentional creative choices or domain-specific descriptions. This nuanced approach ensures that corrections are made only when necessary and appropriate 1.
The Hallucination Corrector works in conjunction with Vectara's widely used Hughes Hallucination Evaluation Model (HHEM). HHEM provides a way to compare responses to source documents and identify if statements are accurate at runtime, scoring answers on a scale from 0 (completely inaccurate) to 1 (perfect accuracy) 2.
By reducing hallucination rates to approximately 0.9% in initial testing, Vectara's solution addresses a critical barrier to enterprise AI adoption, particularly in highly regulated industries such as financial services, healthcare, and law 2.
The Hallucination Corrector offers flexibility in its application. It can automatically use corrected outputs in summaries for end-users, while experts can utilize the full explanation and suggested fixes to refine their models and guardrails. Additionally, the system can flag potential issues in the original summary while offering the corrected version as an optional fix 2.
Amazon Web Services introduces Automated Reasoning checks to tackle AI hallucinations and Model Distillation for creating smaller, efficient AI models, along with multi-agent collaboration features in Amazon Bedrock.
7 Sources
7 Sources
An exploration of AI hallucinations, their causes, and potential consequences across various applications, highlighting the need for vigilance and fact-checking in AI-generated content.
8 Sources
8 Sources
Microsoft introduces a groundbreaking AI correction feature designed to address the issue of AI hallucinations. This development promises to enhance the reliability of AI-generated content across various applications.
2 Sources
2 Sources
AI hallucinations, while often seen as a drawback, offer valuable insights for businesses and healthcare. This article explores the implications and potential benefits of AI hallucinations in various sectors.
2 Sources
2 Sources
Microsoft introduces a new AI-powered tool designed to identify and correct factual errors in AI-generated content. The technology aims to enhance the reliability of AI outputs, but experts warn of potential limitations.
5 Sources
5 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved