Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Report 3008

Associated Incidents

Incident 54158 Report
ChatGPT Reportedly Produced False Court Case Law Presented by Legal Counsel in Court

Loading...
A lawyer used ChatGPT for legal filing. The chatbot cited nonexistent cases it just made up
mashable.com · 2023

Lawyer Steven Schwartz of Levidow, Levidow & Oberman has been practicing law for three decades. Now, one case can completely derail his entire career.

Why? He relied on ChatGPT in his legal filings(opens in a new tab) and the AI chatbot completely manufactured previous cases, which Schwartz cited, out of thin air.

It all starts with the case in question, Mata v. Avianca. According to the New York Times(opens in a new tab), an Avianca(opens in a new tab) customer named Roberto Mata was suing the airline after a serving cart injured his knee during a flight. Avianca attempted to get a judge to dismiss the case. In response, Mata's lawyers objected and submitted a brief filled with a slew of similar court decisions in the past. And that's where ChatGPT came in.

Schwartz, Mata's lawyer who filed the case in state court and then provided legal research once it was transferred to Manhattan federal court, said he used OpenAI's popular chatbot in order to "supplement" his own findings.

ChatGPT provided Schwartz with multiple names of similar cases: Varghese v. China Southern Airlines, Shaboon v. Egyptair, Petersen v. Iran Air, Martinez v. Delta Airlines, Estate of Durden v. KLM Royal Dutch Airlines, and Miller v. United Airlines.

The problem? ChatGPT completely made up all those cases. They do not exist.

Avianca's legal team and the judge assigned to this case soon realized they could not locate any of these court decisions. This led to Schwartz explaining what happened in an affidavit on Thursday. The lawyer had referred to ChatGPT for help with his filing.

According to Schwartz, he was "unaware of the possibility that its content could be false.” The lawyer even provided screenshots to the judge of his interactions with ChatGPT, asking the AI chatbot if one of the cases were real. ChatGPT responded that it was. It even confirmed that the cases could be found in "reputable legal databases." Again, none of them could be found because the cases were all created by the chatbot.

It's important to note that ChatGPT, like all AI chatbots, is a language model trained to follow instructions and provide a user with a response to their prompt. That means, if a user asks ChatGPT for information, it could give that user exactly what they're looking for, even if it's not factual. 

The judge has ordered a hearing next month to "discuss potential sanctions" for Schwartz in response to this “unprecedented circumstance." That circumstance again being a lawyer filing a legal brief using fake court decisions and citations provided to him by ChatGPT.

Read the Source

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd