The Scale of the Problem: 2026 by the Numbers
Data-driven section with specific citations. Pull from Charlotin database, Drug & Device Law monthly counts, Bloomberg Law, and Stateline.
- Total cases: 1,031+ globally, growing at 30–50 per month
- US cases: majority — 518+ since Jan 2025 per Stateline reporting
- Monthly acceleration: Dec 2025 = 51 lawyer cases, Jan 2026 = 36, Feb 2026 (partial) = 33. Source: Drug & Device Law blog’s manual count from the Charlotin database.
- Sanctions range: $1,000 to $86,000 per incident
- Practice areas hit: PI, commercial litigation, family law, bankruptcy, employment, immigration, IP, consumer protection
- Who’s affected: solo practitioners, mid-size firms, Am Law 100 (Gordon Rees, Boies Schiller), even one federal judge
- AI tools involved: ChatGPT (most common), Claude, Gemini, Copilot, and “unspecified AI tools”
For background on how AI hallucinations work in legal contexts, see our guide: 5 AI Hallucination Facts Lawyers Must Know
7 Landmark Cases Every Lawyer Should Know
| Case | Court | Sanction | What Happened | Key Lesson |
|---|---|---|---|---|
| ByoPlanet v. Johansson (Aug 2025) | S.D. Fla. | $86,000 | Repeated, systemic AI misuse across multiple filings despite warnings. Cases dismissed with prejudice. | Largest sanction to date. Court: "A reasonable attorney does not blindly rely on AI." Bad faith + repeated conduct = catastrophic consequences. |
| Fletcher v. Experian (Feb 18, 2026) | 5th Circuit | $2,500 | Published opinion. 16 fabricated quotes + 5 misrepresentations in reply brief. Lawyer denied AI use, then changed story. | Fifth Circuit: dishonesty about AI use triggers harsher penalties. "Had [counsel] accepted responsibility...lesser sanctions." |
| Cassata v. Macrina (Feb 2026) | NY State (Suffolk) | $10,000 | AI-generated citations + plagiarized third-party brief. Judge created a SANCTIONS CHART for AI-related errors. | First judicial sanctions framework specifically for AI errors. Courts are now systematizing penalties. |
| Gordon Rees (2025-2026) | Multiple | Multiple | Am Law 100 firm sanctioned in Jackson Hosp (2025), then ACCUSED AGAIN in Huynh v Redis (Feb 2026). Multiple courts, multiple briefs. | Policies alone don't work. A firm-wide AI policy did not prevent a second incident. Process + tools required. |
| Mostafavi (Sept 2025) | CA 2nd DCA | $10,000 | 21 of 23 quotes in opening brief fabricated by ChatGPT. Lawyer said he "didn't know ChatGPT would add citations." | Largest California state court fine. Published as a warning. Ignorance of AI limitations is not a defense. |
| Morgan & Morgan (2025) | D. Wyo. | $5,000 | 900-lawyer PI firm. Enterprise AI platform still hallucinated. Internal email warning of termination leaked. | Even large firms with AI budgets and policies are vulnerable. Firm had to withdraw motions, pay fees, update policies. |
| Mata v. Avianca (2023) | S.D.N.Y. | $5,000 | The original case. Attorney believed ChatGPT was a "super search engine." 6 fabricated cases submitted. | Established that AI output must be independently verified. Started the entire sanctions trend. |
For a detailed look at how NexLaw compares to Harvey AI on citation accuracy, see our full comparison: Nexlaw v. Harvey AI
What Courts Are Actually Requiring Now
- Federal: No uniform rule. The Fifth Circuit then withdrew a mandatory AI disclosure rule, concluding “existing rules were sufficient.” But the Fletcher opinion shows they’re enforcing aggressively regardless.
- Individual judges: Growing number of standing orders requiring AI disclosure before filing. SDNY, EDTX (Local Rule AT-3), multiple others.
- State courts: Patchwork. Illinois Supreme Court AI policy (2025). California published Mostafavi as a warning. NY Commercial Division proposing AI-specific rules.
- ABA: Formal Opinion (2024) — Rule 1.1 duty of competence requires understanding AI capabilities and limitations. Rule 1.1 Comment 8 on technology competence. Rule 3.3 duty of candor. Rule 5.1 supervisory duties.
- New development: Courts now saying lawyers may have a duty to FLAG AI hallucinations in OPPOSING counsel’s briefs (r/Lawyertalk thread, 130+ comments). This creates a new professional obligation.
- Sanctions escalation trend: From warnings → $1K–5K fines → $10K+ → $86K → calls for disbarment. Bloomberg Law editorial advocates mandatory Congressional reporting.
Why Generic AI Keeps Failing Lawyers
- ChatGPT, Claude, Gemini are text prediction models, not legal databases. They do not access Westlaw, Lexis, or any court system.
- They generate text based on patterns. Legal citation formats (case name, volume, reporter, page) follow predictable patterns that are easy to fabricate.
- The “confidence trap”: AI outputs look indistinguishable from real citations. Even experienced lawyers can’t tell by reading the output alone.
- Stanford/Yale research (Dahl et al., 2024): even legal-specific RAG-based tools hallucinate 17–34% of the time. This means that even some “legal AI” tools beyond ChatGPT carry risk.
- The Fifth Circuit’s practical advice (from Fletcher opinion): “If an LLM’s response to a query seems ‘too good to be true’—that a case or two are unusually helpful or providing a quote that is amazingly on point—it is probably, in fact, too good to be true.”
- The Texas Lawbook analysis of Fletcher: “A carpenter would not use a screwdriver instead of a hammer to drive a nail, and a conscientious lawyer should also not use the wrong tool for the wrong task.”
Why ChatGPT Creates Fake Legal Cases
- ChatGPT is trained on general internet text, not verified legal databases — it has no access to Westlaw, Lexis, or court filing systems.
- Legal citation formats (case name v. case name, volume, reporter, page number) follow highly predictable patterns that are trivially easy for a language model to generate — the format looks real even when the case is invented.
- ChatGPT is designed to be helpful and will generate a plausible-sounding answer rather than say “I don’t know” — researcher Damien Charlotin notes that “the harder your legal argument is to make, the more the model will tend to hallucinate, because they will try to please you.”
- The Mata v. Avianca attorney described ChatGPT as a “super search engine” — this fundamental misunderstanding of what the tool actually does is at the root of most sanctions cases.
For a deeper analysis of why general-purpose AI falls short in legal practice
See our full comparison: Legal tech vs. General AI
The 8-Step AI Compliance Checklist for 2026
- Never use general-purpose AI (ChatGPT, Gemini, Claude) as your primary legal research tool. These are text generators, not legal databases.
- Verify every citation against a primary source — Westlaw, Lexis, a court’s official records, or a citation-backed legal AI platform.
- Check your court’s standing orders and local rules for AI disclosure requirements BEFORE filing. The patchwork is growing fast.
- Supervising attorneys: you are personally liable for AI-generated content you sign. The Fletcher and Cassata courts both sanctioned supervisors.
- Document your verification process. Courts look favorably on good-faith efforts. Keep a verification log for every AI-assisted filing.
- If you discover an error post-filing, disclose and correct IMMEDIATELY. The Fletcher court explicitly said early honesty reduces sanctions. Cover-ups make it catastrophically worse.
- Watch for red flags: the same case cited multiple times, quotes that seem “too perfectly on point,” and any citation you can’t find in 30 seconds on Westlaw or a court site.
- Use legal-specific AI tools with built-in citation verification — not general chatbots. Tools like NexLaw’s NeXa verify every answer against primary legal sources before returning results, eliminating fabricated citations at the source. And even with legal-specific tools, verify independently. Stanford found RAG-based tools still hallucinate 17–34% of the time.
What Citation-Backed Legal AI Actually Looks Like
- Citation-backed research across all 50 states + federal courts
- Every answer verified against primary sources — no hallucinated citations
- Multi-jurisdictional coverage (US, UK, Australia, Canada, Singapore, Malaysia, New Zealand)
- SOC 2 Type II certified
- 3-day free trial, no credit card required
See how NexLaw compares to other legal AI tools
See our full comparison: Nexlaw AI vs. Competitors
Ready to research without the risk?
Ready to research without the risk? NexLaw’s NeXa delivers citation-backed answers verified against primary legal sources. No hallucinated citations. No fabricated case law


