This database tracks legal decisions1
I.e., all documents where the use of AI, whether established or merely alleged, is addressed in more than a passing reference by the court or tribunal.
Notably, this does not cover mere allegations of hallucinations, but only cases where the court or tribunal has explicitly found (or implied) that a party relied on hallucinated content or material.
As an exception, the database also covers some judicial decisions where AI use was alleged but not confirmed. This is a judgment call on my part.
in cases where generative AI produced hallucinated content – typically fake citations, but also other types of AI-generated arguments. It does not track the (necessarily wider) universe of all fake citations or use of AI in court filings.
While seeking to be exhaustive (6 cases identified so far), it is a work in progress and will expand as new examples emerge. This database has been featured in news media, and indeed in several decisions dealing with hallucinated material.2
Examples of media coverage include:
- M. Hiltzik, AI 'hallucinations' are a growing problem for the legal profession (LA Times, 22 May 2025)
- E. Volokh, "AI Hallucination Cases," from Courts All Over the World (Volokh Conspiracy, 18 May 2025)
- J-.M. Manach, "Il génère des plaidoiries par IA, et en recense 160 ayant « halluciné » depuis 2023" (Next, 1 July 2025)
- J. Koebler & J. Roscoe, "18 Lawyers Caught Using AI Explain Why They Did It (404 Media, 30 September 2025)
Based on this database, I have developped an automated reference checker that also detects hallucinations: PelAIkan. Check the Reports
in the database for examples, and reach out to me for a demo !
For weekly takes on cases like these, and what they mean for legal practice, subscribe to Artificial Authority.
| Case | Court / Jurisdiction | Date ▼ | Party Using AI | AI Tool ⓘ | Nature of Hallucination | Outcome / Sanction | Monetary Penalty | Details | Report(s) |
|---|---|---|---|---|---|---|---|---|---|
| United States v. Farris | CA 6th Cir. (USA) | 3 April 2026 | Lawyer | Westlaw CoCounsel |
Fabricated
Doctrinal Work
(1)
False Quotes
Case Law
(2)
|
Counsel disqualified with no compensation for time served; Briefs locked; Bar Referral; Notice of Opinion | — | — | |
|
Source: Robert Freund
|
|||||||||
| Heimkes v. Fairhope Motorcoach Resort Condominium Owners Association, Inc. | S.D. Alabama (USA) | 31 March 2026 | Lawyer | Cocounsel (Westlaw) |
Fabricated
Case Law
(1)
Misrepresented
Case Law
(1)
Outdated Advice
Overturned Case Law
(1)
|
Reprimand; Order to file order in subsequent cases; Bar Referral; Adverse Costs Order | 55597 | — | |
| CVTEK, LLC (B-423943; B-423943.2) | GAO (USA) | 12 February 2026 | Lawyer | Westlaw CoCounsel |
Fabricated
Case Law
(1)
Misrepresented
Case Law
(1)
|
Warning | — | — | |
| Anthony C. Hill v. Workday, Inc. | N.D. California (USA) | 5 September 2025 | Lawyer | CoCounsel |
Fabricated
Case Law
(1)
Misrepresented
Case Law
(1)
|
Order to circulate decision in law firm; CLE | — | — | |
|
"In her declaration, Attorney Cervantes described the erroneous citation as her own inadvertent error and further implied that Westlaw’s tool may have “glitched” by erroneously producing the citation. [Dkt. 31 at ¶ 4]. Specifically, she indicates that the copy citation tool mayhave inadvertently copied the wrong citation information. Yet, that explanation seems highly improbable. The citation was incorrect in every respect and none of the sub-parts of the citation match each other, including the case docket number, reporter, pincite, court information, and date information. The likelihood of every component of a citation being simultaneously wrong due to a software malfunction in transcribing from a “correct” citation is, in the Court’s view, statistically improbable (particularly given Westlaw’s representations as to the accuracy of its tools and the apparent resources Westlaw has devoted to promoting the reputation of this tool). If one set of numbers had been transposed, or if the date were wrong, such transcription errors might be explicable. However, the fact that the party names, docket number, Westlaw citation, date, and court all failed to match, resulting in a mashup citation, cannot credibly be attributed to a software error, let alone a mistranscription or copying mistake. Rather,this Frankensteinian legal citation, stitched together from mismatched party names, docket numbers, dates, and courts, bears the hallmarks of an AI-generated hallucination, as documented in numerous published opinions and reports." |
|||||||||
| Lacey v. State Farm General Insurance | C.D. Cal (USA) | 6 May 2025 | Lawyer | CoCounsel, Westlaw Precision, Google Gemini |
Fabricated
Case Law
(2)
False Quotes
Case Law
(4)
|
Striking of briefs; denial of requested discovery relief; Large monetary sanctions jointly and severally against the two law firms | 31100 USD | — | |
AI UseCounsel used CoCounsel, Westlaw’s AI tools, and Google Gemini to generate a legal outline for a discovery-related supplemental brief. The outline contained hallucinated citations and quotations, which were incorporated into the filed brief by colleagues at both Ellis George and K&L Gates. No one verified the content before filing. After the Special Master flagged two issues, counsel refiled a revised brief—but it still included six AI-generated hallucinations and did not disclose AI use until ordered to respond. Hallucination DetailsAt least two cases did not exist at all, including a fabricated quotation attributed to Booth v. Allstate Ins. Co., 198 Cal.App.3d 1357 (1989). Misquoted or fabricated quotes attributed to National Steel Products Co. v. Superior Court, 164 Cal.App.3d 476 (1985). Several additional misquotes and garbled citations across three submitted versions of the brief. Revised versions attempted to silently “fix” errors without disclosing their origin in AI output. Ruling/SanctionThe Special Master (Judge Wilner) struck all versions of Plaintiff’s supplemental brief, denied the requested discovery relief, and imposed:
Key Judicial ReasoningThe submission and re-submission of AI-generated material without verification, especially after warning signs were raised, was deemed reckless and improper. The court emphasized that undisclosed AI use that results in fabricated law undermines judicial integrity. While individual attorneys were spared, the firms were sanctioned for systemic failure in verification and supervision. The Special Master underscored that the materials nearly made it into a judicial order, calling that prospect “scary” and demanding “strong deterrence.” |
|||||||||
| Williams v. Capital One Bank | D. DC (USA) | 18 March 2025 | Pro Se Litigant | CoCounsel |
Fabricated
Case Law
(1)
Misrepresented
Case Law
(1)
|
Case dismissed with prejudice for failure to state a claim. No monetary sanction imposed, but the court issued a formal warning | — | — | |
AI UseWhile not formally admitted, Plaintiff’s opposition brief referred to “legal generative AI program CoCounsel,” and the court noted that the structure and citation pattern were consistent with AI-generated output. Capital One was unable to verify several case citations, prompting the court to scrutinize the submission. Hallucination DetailsAt least one case was fully fabricated, and another was a real case misattributed to the wrong jurisdiction and reporter. The court emphasized that it could not determine whether the mis-citations were the result of confusion, poor research, or hallucinated AI output—but the burden rested with the party filing them. Ruling/SanctionThe court dismissed the complaint with prejudice, noting Plaintiff had already filed and withdrawn a prior version and had had full opportunity to amend. Though it did not impose monetary sanctions, it issued a strong warning and directed Plaintiff to notify other courts where he had similar pending cases if any filings included erroneous AI-generated citations. |
|||||||||