Upcoming Webinar : Security Foundations for Agentic AI - Register Now !

OWASP LLM09: Misinformation Risks in Production AI Applications

Generative AI has fundamentally changed how software is built and how information is consumed.  As organizations rely more on Large Language Models (LLMs), a new risk has become increasingly clear: AI systems can generate confident but incorrect outputs that users trust and act on.

In the OWASP Top 10 2025 for LLM Applications,  this risk is formally identified as LLM09: Misinformation.

LLM-generated misinformation can lead to real security and business impact, including compliance violations, faulty decisions, and downstream system failures. This blog explains what LLM09 means in practice, why it occurs, and how security teams can reduce this risk effectively.

What OWASP Means by LLM09 – Misinformation

One common misunderstanding is to equate LLM misinformation with disinformation. The two are very different.

Disinformation involves intent, deliberately spreading false information to mislead. LLM09, as defined by OWASP, focuses on something far more subtle and dangerous: models generating false, misleading, or nonsensical content that appears authoritative and credible.

In most cases, the model is not behaving maliciously. It is behaving exactly as designed.

LLMs are probabilistic systems. They predict the next most likely word based on patterns learned during training. They do not verify facts, reason about truth, or understand consequences. When an LLM responds with absolute confidence but incorrect information, it creates what many teams discover too late, a trust trap. Users assume accuracy because the answer sounds correct, and systems act on outputs that were never validated.

That gap between confidence and correctness is where misinformation becomes operationally dangerous.

Why LLMs Produce Convincing Falsehoods

To control misinformation, teams need to understand where it comes from. These failures are not random; they follow predictable patterns.

1. Hallucination: The Primary Engine

Hallucination is the most common driver of misinformation. When an LLM lacks sufficient information, encounters ambiguity, or is pushed outside its training scope, it rarely admits uncertainty. Thus, filling the gap with a statistically plausible answer.

The real risk lies in “how” these answers are delivered. Hallucinated outputs often use the correct tone, structure, and terminology like legal language that sounds precise, medical explanations that feel clinical, or code that looks syntactically valid. Briefly, there is nothing that signals the information is fabricated.

2. Training Data Limitations

An LLM can only reflect what it has seen. If its training data contains outdated material, biased viewpoints, or factual gaps, those weaknesses are reproduced with confidence. The model does not know that a regulation has changed; a vulnerability has been patched, or a library has been deprecated, unless it is explicitly grounded in up-to-date sources.

3. Lack of Grounding to a Source of Truth

When an LLM operates without access to verified documents or live data, it relies entirely on internal model weights. In domains like cybersecurity, finance, or law, where accuracy depends on current information, this becomes a serious liability. The moment training ends; the knowledge begins to age.

The Human Factor: Over-Reliance as a Risk Multiplier

Over-reliance used to be listed as its own OWASP category. In 2025, it was folded into LLM09 for a reason: misinformation rarely causes damage on its own. It becomes a security incident when people trust it without question.

This shows up in familiar ways:

  • A developer copies AI-generated code straight into production.
  • A lawyer cites a case that never existed.
  • A customer support chatbot commits the company to a policy it does not offer.

Because LLMs respond fluently and confidently, users often skip verification. What starts as a probabilistic suggestion quickly becomes a perceived source of truth. At that point, failure is no longer theoretical; it is operational.

How Misinformation Appears in Real Systems

Misinformation does not surface randomly. It tends to cluster around specific use cases.

1. Package Hallucination and Supply-Chain Risk

One of the most serious risks for application security teams comes from AI-assisted development. LLMs regularly suggest software packages or libraries that sound legitimate but do not exist.

Attackers have learned to exploit this behavior. They monitor common hallucinated package names, register them on public repositories, and seed them with malicious code. When a developer installs the package suggested by the model, the compromise happens silently, inside the build pipeline.

Explore the key steps to prevent LLM supply chain risks.

2. Customer-Facing Accuracy Failures

When LLMs are embedded in chatbots or virtual agents, they effectively speak on behalf of the organization. A hallucinated refund policy, incorrect safety guidance, or fabricated entitlement can quickly escalate into legal disputes and regulatory scrutiny.

3. Fabricated Expertise in High-Stakes Domains

In healthcare and legal environments, being “almost right” is not acceptable. LLMs have been shown to invent legal precedents, misinterpret clinical guidance, or suggest treatments based on linguistic patterns rather than evidence. These failures carry real risk, even when no attacker is involved.

Mitigating LLM09: Going Beyond Better Prompts

Reducing misinformation in LLM-based systems requires layered controls rather than a single fix. The goal is to limit hallucinations, detect errors early, and prevent unchecked outputs from being treated as authoritative.

1. Ground Responses Using Retrieval-Augmented Generation (RAG)

Retrieval-Augmented Generation helps control misinformation by anchoring responses to trusted data sources. Using curated documents or internal knowledge bases ensures outputs remain aligned with verified information, reducing the likelihood of unsupported or outdated claims.

2. Improve Output Quality Through Fine-Tuning

Fine-tuning and embedding-based approaches help align models with specific domains where accuracy matters. Techniques such as parameter-efficient tuning and structured reasoning prompts improve relevance and consistency, lowering the likelihood of misleading outputs in specialized use cases.

3. Enforce Cross-Verification and Human Review

LLM outputs should not be accepted at face value in high-impact scenarios. Cross-checking responses against reliable sources and involving trained human reviewers helps catch inaccuracies before they propagate. Human oversight is especially important where legal, medical, financial, or security decisions are involved.

4. Apply Automated Validation Controls

Automated validation mechanisms provide scalable protection by checking outputs against predefined rules, constraints, or reference data. In sensitive environments, these controls should flag or block responses that fail accuracy or safety checks.

5. Communicate Risks and Limitations Clearly

Users must understand that LLMs can produce incorrect information. Clearly communicating model limitations and the potential for misinformation reduces blind trust and encourages responsible use of AI-generated content.

6. Maintain Secure Coding Practices

When LLMs assist with software development, their output should be treated as untrusted input. Standard code reviews, dependency checks, and security testing help prevent vulnerabilities introduced by incorrect or hallucinated code suggestions.

7. Design Interfaces That Discourage Over-Reliance

User interfaces and APIs should make it clear when content is AI-generated, highlight accuracy limitations, and define intended use cases. Thoughtful design reduces the risk of users mistaking probabilistic outputs for guaranteed truth.

8. Train Users to Evaluate AI Output Critically

Training users on LLM limitations and verification practices is essential. In domain-specific contexts, targeted education helps users assess whether an output is reasonable, complete, and suitable for the task at hand.

Indusface
Indusface

Indusface is a leading application security SaaS company that secures critical Web, Mobile, and API applications of 5000+ global customers using its award-winning fully managed platform that integrates web application scanner, web application firewall, DDoS & BOT Mitigation, CDN, and threat intelligence engine.

Frequently Asked Questions (FAQs)

Why does OWASP classify misinformation as a top risk for LLM applications?

OWASP identifies misinformation as a top risk because LLM outputs often appear authoritative and are increasingly integrated into automated workflows. When incorrect responses are consumed without validation, they can trigger compliance violations, faulty decisions, and downstream system failures.

Why do Large Language Models generate confident but incorrect answers? +

LLMs are probabilistic systems that predict the next likely word based on training patterns, not factual verification. When faced with ambiguity, missing data, or outdated knowledge, they tend to produce plausible responses rather than admitting uncertainty. This gap between confidence and correctness is the core driver of misinformation.

How does LLM misinformation become a real security risk? +

Misinformation becomes dangerous when AI-generated outputs are treated as trusted inputs. This includes scenarios such as developers deploying hallucinated code, chatbots committing to incorrect policies, or AI systems suggesting non-existent software packages that attackers later weaponize. The impact is operational, not theoretical.

What are common real-world examples of LLM09 misinformation? +

Common examples include hallucinated software packages that introduce supply-chain risk, customer-facing chatbots providing incorrect legal or financial information, and AI systems fabricating expertise in regulated domains like healthcare or law. These failures often occur without any attacker involvement but still result in serious consequences.

How can organizations reduce the risk of LLM-generated misinformation? +

Reducing LLM09 risk requires layered controls such as grounding responses with Retrieval-Augmented Generation (RAG), applying fine-tuning for domain accuracy, enforcing human review in high-impact use cases, and validating outputs automatically. Treating LLM outputs as untrusted by default is critical for maintaining system integrity.

Join 51000+ Security Leaders

Get weekly tips on blocking ransomware, DDoS and bot attacks and Zero-day threats.

We're committed to your privacy. indusface uses the information you provide to us to contact you about our relevant content, products, and services. You may unsubscribe from these communications at any time. For more information, check out our Privacy Policy.

AppTrana

Fully Managed SaaS-Based Web Application Security Solution

Get free access to Integrated Application Scanner, Web Application Firewall, DDoS & Bot Mitigation, and CDN for 14 days

Get Started for Free Request a Demo

Gartner

Indusface is the only cloud WAAP (WAF) vendor with 100% customer recommendation for 4 consecutive years.

A Customers’ Choice for 2024, 2023 and 2022 - Gartner® Peer Insights™

The reviews and ratings are in!