When AI Hallucinates, Who’s Liable?

AI hallucinations can cause financial, legal, and operational harm — but traditional liability laws weren’t built for probabilistic AI systems. Here’s how global standards, regulators, and governance frameworks are redefining responsibility.

Viktorija Isic

|

AI & Ethics

|

September 23, 2025

Listen to this article

0:00/1:34

Introduction: AI Capability Has Outrun Our Accountability Structures

AI systems now summarize medical data, draft legal arguments, generate financial insights, and guide enterprise decision-making. Yet the same systems can also fabricate information with fluency and confidence. These false outputs — known as hallucinations — are not random glitches but structural features of generative models.

As AI becomes embedded in high-stakes environments, a critical governance question emerges:

When AI hallucinates — who is responsible?

Developers? Organizations deploying the technology? The individuals relying on it?

Traditional liability frameworks struggle to answer these questions because AI breaks the assumptions that legal systems were built on. Below is a clear, research-grounded breakdown of how accountability is evolving in the era of AI hallucination risk.

1. What Exactly Is an AI Hallucination? A Technical and Ethical Definition

AI hallucinations occur when a model produces information that is:

  • fabricated

  • inaccurate

  • unsupported by training data

  • expressed with unwarranted certainty

According to Nature Machine Intelligence (2023), hallucinations appear in 15–20% of outputs depending on the reasoning task — even in advanced models (Nature Machine Intelligence, 2023).

Hallucinations matter because users often assume the output is factual, especially in domains like finance, healthcare, and law where information precision is critical.

Ethically, a hallucination becomes harmful when a reasonable user depends on the output to make decisions impacting rights, safety, or financial well-being.

2. Why Liability Is Unclear: AI Breaks Traditional Causality

Traditional liability frameworks rely on:

  • a clear human actor

  • identifiable causation

  • predictable system behavior

  • traceable responsibility

AI systems undermine each of these.

AI introduces:

  • Probabilistic outputs, not deterministic logic

  • Opaque reasoning pathways (“black-box” inference)

  • Distributed responsibility across developers, organizations, and users

  • Unpredictable errors that cannot be fully eliminated

The OECD notes that modern AI creates conditions of “diffuse causality,” making it difficult to attribute fault to a single actor (OECD, 2023).

For organizations, this means that assuming the model is solely at fault is no longer defensible — regulators expect proactive governance.

3. Real-World Harm: How Hallucinations Become Liability Events

AI hallucinations have already produced legally and financially significant failures.

Legal Sector: Fabricated Case Law

A New York lawyer was sanctioned after submitting briefs containing hallucinated case citations generated by a language model (Sneed, 2023).

Healthcare: Unsafe Clinical Recommendations

Stanford researchers found that generative models produced incorrect or unsafe medical guidance in up to 36% of tested scenarios (Stanford HAI, 2023).

Finance: Reputational, Fiduciary, and Regulatory Exposure

Hallucinated financial insights or analysis increase risk across:

  • fiduciary obligations

  • SEC compliance

  • investment decision accuracy

  • audit defensibility

  • consumer protection expectations

Errors can constitute misrepresentation, even if produced by a model.

Research & Education: Fabricated Sources

AI-invented citations undermine academic credibility and institutional trust.

From law to healthcare to capital markets, hallucinations already carry material consequences — and regulators are responding.

4. Who Should Be Liable?

The Emerging Shared-Responsibility Model**

Global governance bodies — including the EU, NIST, UNESCO, OECD, and WEF — increasingly converge on a shared liability structure across three actors.

A. Developers: Liability for Safety, Testing, Transparency & Risk Disclosure

Developers are accountable for:

  • documenting model limitations

  • conducting safety evaluations

  • red-teaming and stress-testing

  • improving reliability

  • ensuring safe scaling

  • providing transparency on training data and risks

Under the EU Artificial Intelligence Act, developers of high-risk systems face legal obligations and penalties for inadequate testing or documentation (European Commission, 2024).

B. Deploying Organizations: Primary Responsibility for Governance & Oversight

Organizations integrating AI into workflows bear the strongest responsibility because they control context, use cases, and downstream impact.

NIST’s AI Risk Management Framework emphasizes that organizations must implement:

  • model-risk governance

  • human oversight protocols

  • documented evaluation processes

  • continuous monitoring

  • incident reporting

  • domain-appropriate safeguards

(NIST, 2023).

MIT Technology Review adds that AI liability is shifting toward a cybersecurity-style model, where organizations must prove adequate controls and due diligence (Metz, 2023).

C. End-Users: Limited but Real Responsibility

Users are expected to:

  • verify high-risk outputs

  • follow organizational policies

  • avoid misuse

  • disclose AI involvement when required

However, frameworks emphasize users cannot carry the primary burden — they lack the technical context to assess model reliability (UNESCO, 2021).

5. The Global Liability Landscape: Regulation Is Accelerating

European Union (EU AI Act)

The most advanced regulatory regime, introducing:

  • strict documentation

  • testing requirements

  • transparency obligations

  • enforcement penalties

  • high-risk system classifications

(European Commission, 2024).

United States (NIST, FTC, SEC)

The U.S. uses a soft-law approach:

  • NIST AI RMF

  • FTC unfair-practices enforcement

  • SEC expectations around model governance

  • White House Executive Order on AI

Regulators expect organizations to implement governance even without explicit statutes.

Global Standards (UNESCO, OECD, WEF)

Ethical AI frameworks call for:

  • traceability

  • human oversight

  • transparency

  • impact assessment

  • continuous monitoring

(UNESCO, 2021; OECD, 2023; World Economic Forum, 2023).

Together, these form an international shift toward mandatory organizational accountability.

6. Reducing Liability: What Organizations Must Do Immediately

Leaders adopting AI should act now to reduce exposure:

Implement enterprise-grade model-risk management (MRM)

Borrowed from finance — essential for audit and defensibility.

Require human verification for high-context tasks

Especially in healthcare, legal, or financial contexts.

Stress-test models continuously

Hallucination frequencies and patterns evolve over time.

Create clear documentation and audit trails

Organizations must be able to demonstrate responsible use.

Train staff on responsible AI use

Governance fails without human understanding.

Establish AI incident reporting pathways

Hallucinations must be logged, triaged, and reviewed.

Conclusion: Hallucinations Aren’t the Risk — Unprepared Systems Are

AI hallucinations are inevitable. But preventable harm is not. The greatest liability exposure today does not come from AI itself — but from organizations deploying it without:

  • governance

  • oversight

  • traceability

  • testing

  • documentation

  • ethical guardrails

The next decade will reward leaders who treat AI governance not as compliance, but as strategy, resilience, and competitive advantage. If you’re implementing AI in finance, healthcare, legal, policy, or other regulated industries, you need strong governance — not guesswork. Subscribe to my weekly insights for executive-level guidance on AI ethics, governance, and risk. Request a strategy consultation if your organization needs help building or assessing AI governance structures. Let’s build systems that scale — safely, ethically, and intelligently.

References

  1. European Union. (2024). Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence (Artificial Intelligence Act).
    https://eur-lex.europa.eu/eli/reg/2024/1689/oj/eng

  2. Jones, N. (2025). AI hallucinations can’t be stopped — but these techniques can limit their damage. Nature.
    https://www.nature.com/articles/d41586-025-00068-5

  3. MIT Sloan School of Management. (2023, August 28). The legal issues presented by generative AI.
    https://mitsloan.mit.edu/ideas-made-to-matter/legal-issues-presented-generative-ai

  4. National Institute of Standards and Technology. (2023). AI Risk Management Framework (AI RMF 1.0).
    https://www.nist.gov/itl/ai-risk-management-framework

  5. Organisation for Economic Co-operation and Development. (2019). OECD AI principles.
    https://oecd.ai/en/ai-principles

  6. Stanford Institute for Human-Centered Artificial Intelligence. (2025). Holistic evaluation of large language models for medical applications.
    https://hai.stanford.edu/news/holistic-evaluation-large-language-models-medical-applications

  7. Stanford Medicine. (2025, April 8). Evaluating AI in context: Which LLM is best for real health care needs?
    https://med.stanford.edu/news/insights/2025/04/ai-artificial-intelligence-evaluation-algorithm.html

  8. Sneed, T. (2025, May 31). US lawyer sanctioned after being caught using ChatGPT for court brief. The Guardian.
    https://www.theguardian.com/us-news/2025/may/31/utah-lawyer-chatgpt-ai-court-brief

  9. UNESCO. (2022). Recommendation on the ethics of artificial intelligence.
    https://www.unesco.org/en/articles/recommendation-ethics-artificial-intelligence

  10. World Economic Forum. (2021). The AI governance journey: Development and opportunities.
    https://www3.weforum.org/docs/WEF_The%20AI_Governance_Journey_Development_and_Opportunities_2021.pdf

Want more insights like this? 

Subscribe to my newsletter or follow me on LinkedIn for fresh perspectives on leadership, ethics, and AI

Subscribe to my newsletter