Close Menu
TechurzTechurz

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    I asked Google Finance’s AI chatbot what stocks to buy – and its answer surprised me

    August 28, 2025

    Intel has received $5.7 billion under Trump’s investment deal

    August 28, 2025

    This Qi2 battery pack from Anker just made wireless charging essential for me

    August 28, 2025
    Facebook X (Twitter) Instagram
    Trending
    • I asked Google Finance’s AI chatbot what stocks to buy – and its answer surprised me
    • Intel has received $5.7 billion under Trump’s investment deal
    • This Qi2 battery pack from Anker just made wireless charging essential for me
    • Bob Odenkirk’s ‘Nobody 2’ Gets Streaming Date, Report Says
    • Unravelling 5G Complexity: Engaging Students with TIMS-Powered Hands-on Education
    • Scientists Are Flocking to Bluesky
    • MathGPT, the ‘cheat-proof’ AI tutor and teaching assistant, expands to over 50 institutions
    • The Download: Google’s AI energy use, and the AI Hype Index
    Facebook X (Twitter) Instagram Pinterest Vimeo
    TechurzTechurz
    • Home
    • AI
    • Apps
    • News
    • Guides
    • Opinion
    • Reviews
    • Security
    • Startups
    TechurzTechurz
    Home»AI»Guardian agents: New approach could reduce AI hallucinations to below 1%
    AI

    Guardian agents: New approach could reduce AI hallucinations to below 1%

    TechurzBy TechurzMay 13, 2025No Comments7 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Guardian agents: New approach could reduce AI hallucinations to below 1%
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More

    Hallucination is a risk that limits the real-world deployment of enterprise AI.

    Many organizations have attempted to solve the challenge of hallucination reduction with various approaches, each with varying degrees of success. Among the many vendors that have been working for the last several years to reduce the risk is Vectara. The company got its start as an early pioneer in grounded retrieval, which is better known today by the acronym Retrieval Augmented Generation (RAG). An early promise of RAG was that it could help reduce hallucinations by sourcing information from provided content.

    While RAG is helpful as a hallucination reduction approach, hallucinations still occur even with RAG. Among existing industry solutions most  solutions focus on detecting hallucinations or implementing preventative guardrails, Vectara has unveiled a fundamentally different approach: automatically identifying, explaining and correcting AI hallucinations through what it calls guardian agents inside of a new service called the Vectara Hallucination Corrector.

    The guardian agents are functionally software components that monitor and take protective actions within AI workflows. Instead of just applying rules inside of an LLM, the promise of guardian agents is to apply corrective measures in an agentic AI approach that improves workflows. Vectara’s approach makes surgical corrections while preserving the overall content and providing detailed explanations of what was changed and why.

    The approach appears to deliver meaningful results. According to Vectara, the system can reduce hallucination rates for smaller language models under 7 billion parameters, to less than 1%.

    “As enterprises are implementing more agentic workflows, we all know that hallucinations are still an issue with LLMs and how that is going to exponentially amplify the negative impact of making mistakes in an agentic workflow is kind of scary for enterprises,” Eva Nahari, chief product officer at Vectara told VentureBeat in an exclusive interview. “So what we have set out as a continuation of our mission to build out trusted AI and enable the full potential of gen AI for enterprise… is this new track of releasing guardian agents.”

    The enterprise AI hallucination detection landscape

    Every enterprise wants to have accurate AI, that’s not a surprise. It’s also no surprise that there are many different options for reducing hallucinations.

    RAG approaches help to reduce hallucinations by providing grounded responses from content but can still yield inaccurate results. One of the more interesting implementations of RAG is one from the Mayo Clinic  which uses a ‘reverse RAG‘ approach to limit hallucinations.

    Improving data quality as well as how vector data embeddings are created is another approach to improving accuracy. Among the many vendors working on that approach is database vendor MongoDB which recently acquired advanced embedding and retrieval model vendor Voyage AI.

    Guardrails, which are available from many vendors including Nvidia and AWS among others, help to detect risky outputs and can help with accuracy in some cases. IBM actually has a set of its Granite open-source models known as Granite Guardian that directly integrate guardrails as a series of fine-tuning instructions, to reduce risky outputs.

    Using reasoning to validate output is another potential solution. AWS claims that its Bedrock Automated Reasoning approach catches 100% of hallucinations, though that claim is difficult to validate.

    Startup Oumi offers another approach, validating claims made by AI on a sentence by sentence basis by validating source materials with an open-source technology called HallOumi.

    How the guardian agent approach is different

    While there is merit to all the other approaches to hallucination reduction, Vectara claims its approach is different.

    Rather than just identifying if a hallucination is present and then either flagging or rejecting the content, the guardian agent approach actually corrects the issue. Nahari emphasized that the guardian agent takes action. 

    “It’s not just a learning on something,” she said. “It’s taking an action on behalf of someone, and that makes it an agent.”

    The technical mechanics of guardian agents

    The guardian agent is a multi-stage pipeline rather than a single model.

    Suleman Kazi, machine learning tech lead at Vectara told VentureBeat that the system comprises three key components: a generative model, a hallucination detection model and a hallucination correction model. This agentic workflow allows for dynamic guardrailing of AI applications, addressing a critical concern for enterprises hesitant to fully embrace generative AI technologies.

    Rather than wholesale elimination of potentially problematic outputs, the system can make minimal, precise adjustments to specific terms or phrases. Here’s how it works:

    1. A primary LLM generates a response
    2. Vectara’s hallucination detection model (Hughes Hallucination Evaluation Model) identifies potential hallucinations
    3. If hallucinations are detected above a certain threshold, the correction agent activates
    4. The correction agent makes minimal, precise changes to fix inaccuracies while preserving the rest of the content
    5. The system provides detailed explanations of what was hallucinated and why

    Why nuance matters for hallucination detection

    The nuanced correction capabilities are critically important. Understanding the context of the query and source materials can make the difference between an answer being accurate or being a hallucination.

    When discussing the nuances of hallucination correction, Kazi provided a specific example to illustrate why blanket hallucination correction isn’t always appropriate. He described a scenario where an AI is processing a science fiction book that describes the sky as red, instead of the typical blue. In this context, a rigid hallucination correction system might automatically “correct” the red sky to blue, which would be incorrect for the creative context of a science fiction narrative. 

    The example was used to demonstrate that hallucination correction needs contextual understanding. Not every deviation from expected information is a true hallucination – some are intentional creative choices or domain-specific descriptions. This highlights the complexity of developing an AI system that can distinguish between genuine errors and purposeful variations in language and description.

    Alongside its guardian agent, Vectara is releasing HCMBench, an open-source evaluation toolkit for hallucination correction models.

    This benchmark provides standardized ways to evaluate how well different approaches correct hallucinations. The goal of the benchmark is to help the community at large, as well as to help enable enterprises to evaluate hallucination correction claims accuracy, including those from Vectara. The toolkit supports multiple metrics including HHEM, Minicheck, AXCEL and FACTSJudge, providing comprehensive evaluation of hallucination correction effectiveness.

    “If the community at large wants to develop their own correction models, they can use that benchmark as an evaluation data set to improve their models,” Kazi said.

    What this means for enterprises

    For enterprises navigating the risks of AI hallucinations, Vectara’s approach represents a significant shift in strategy. 

    Instead of just implementing detection systems or abandoning AI in high-risk use cases, companies can now consider a middle path: implementing correction capabilities. The guardian agent approach also aligns with the trend toward more complex, multi-step AI workflows.

    Enterprises looking to implement these approaches should consider:

    1. Evaluating where hallucination risks are most critical in their AI implementations.
    2. Considering guardian agents for high-value, high-risk workflows where accuracy is paramount.
    3. Maintaining human oversight capabilities alongside automated correction.
    4. Leveraging benchmarks like HCMBench to evaluate hallucination correction capabilities.

    With hallucination correction technologies maturing, enterprises may soon be able to deploy AI in previously restricted use cases while maintaining the accuracy standards required for critical business operations.

    Daily insights on business use cases with VB Daily

    If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

    Read our Privacy Policy

    Thanks for subscribing. Check out more VB newsletters here.

    An error occured.

    agents approach Guardian hallucinations reduce
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticlePanasonic’s new Lumix S1 II mirrorless camera is a video powerhouse with some serious upgrades, but it’s pricey for a Lumix
    Next Article nubia Z70S Ultra reaches international markets, nubia Pad Pro follows next month
    Techurz
    • Website

    Related Posts

    AI

    I asked Google Finance’s AI chatbot what stocks to buy – and its answer surprised me

    August 28, 2025
    AI

    Unravelling 5G Complexity: Engaging Students with TIMS-Powered Hands-on Education

    August 28, 2025
    AI

    The Download: Google’s AI energy use, and the AI Hype Index

    August 28, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    You Can Now Get Starlink for $15-Per-Month in New York, but There’s a Catch

    July 11, 20257 Views

    Non-US businesses want to cut back on using US cloud systems

    June 2, 20257 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    You Can Now Get Starlink for $15-Per-Month in New York, but There’s a Catch

    July 11, 20257 Views

    Non-US businesses want to cut back on using US cloud systems

    June 2, 20257 Views
    Our Picks

    I asked Google Finance’s AI chatbot what stocks to buy – and its answer surprised me

    August 28, 2025

    Intel has received $5.7 billion under Trump’s investment deal

    August 28, 2025

    This Qi2 battery pack from Anker just made wireless charging essential for me

    August 28, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and Conditions
    • Disclaimer
    © 2025 techurz. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.