Lawyers Sanctioned After ChatGPT Hallucinates Fake Cases
Key Points
- Multiple lawyers sanctioned for submitting AI-generated fake case citations
- Judges across jurisdictions began requiring disclosure of AI tool usage
- Bar associations issued new ethics guidelines for AI in legal practice
- ChatGPT confidently generated non-existent cases with realistic citations
- Some courts banned AI-generated content in filings entirely
A New York lawyer was sanctioned in June 2023 after submitting a court brief containing fake legal citations generated by ChatGPT. The case became a cautionary tale about AI reliability in professional settings.
A lawyer used ChatGPT to write a court filing and it made up fake court cases. The judge fined the lawyers and it became a warning about trusting AI too much.
Sides
Critics
No critics identified
Defenders
No defenders identified
Neutral
Acknowledged hallucination risks while noting ChatGPT is not designed for legal research
Noise Level
Forecast
AI disclosure requirements in courts will become standard nationwide. Legal AI tools will need to implement citation verification as a core feature.
Based on current signals. Events may develop differently.
Timeline
Judge sanctions lawyers and orders apology letters
Court imposes $5,000 fine and mandates apology to judges cited in fictitious cases
Lawyer submits ChatGPT-generated brief with fake case citations
Steven Schwartz used ChatGPT to research cases that turned out to be entirely fabricated