Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
発見する
投稿する
  • ようこそAIIDへ
  • インシデントを発見
  • 空間ビュー
  • テーブル表示
  • リスト表示
  • 組織
  • 分類法
  • インシデントレポートを投稿
  • 投稿ランキング
  • ブログ
  • AIニュースダイジェスト
  • リスクチェックリスト
  • おまかせ表示
  • サインアップ
閉じる
発見する
投稿する
  • ようこそAIIDへ
  • インシデントを発見
  • 空間ビュー
  • テーブル表示
  • リスト表示
  • 組織
  • 分類法
  • インシデントレポートを投稿
  • 投稿ランキング
  • ブログ
  • AIニュースダイジェスト
  • リスクチェックリスト
  • おまかせ表示
  • サインアップ
閉じる

レポート 3031

関連インシデント

インシデント 54158 Report
ChatGPT Reportedly Produced False Court Case Law Presented by Legal Counsel in Court

Loading...
New York Lawyer Caught Using ChatGPT After Citing Cases That Don't Exist
iflscience.com · 2023

A lawyer in New York has found himself in trouble with a judge after he submitted legal research which had been created by artificial intelligence (AI) chatbot ChatGPT.

During a case of an airline being sued over an alleged personal injury, lawyers for the plaintiff filed a brief containing several cases to be used as legal precedent. Unfortunately, as later admitted in an affidavit, the following cases were "found to be nonexistent" by the court:

Varghese v. China Southern Airlines Co Ltd, 925 F.3d 1339 (11th Cir. 2019) Shaboon v. Egyptair 2013 IL App (1st) 111279-U (Il App. Ct. 2013)

Petersen v. Iran Air 905 F. Supp 2d 121 (D.D.C. 2012)

Martinez v. Delta Airlines, Inc, 2019 WL 4639462 (Tex. App. Sept. 25, 2019)

Estate of Durden v. KLM Royal Dutch Airlines, 2017 WL 2418825 (Ga. Ct. App. June 5, 2017)

Miller v. United Airlines, Inc, 174 F.3d 366 (2d Cir. 1999)

The "research" was compiled by lawyer Steven A. Schwartz, an attorney with over 30 years of experience according to the BBC. Schwartz said in the affidavit that he had not used ChatGPT for legal research before and was "unaware of the possibility that its content could be false". 

Screenshots in the affidavit show the lawyer asking the chatbot "is varghese a real case", to which the chatbot responded "yes". When asked for sources, it told the lawyer that the case could be found "on legal research databases such as Westlaw and LexisNexis". When asked "are the other cases you provided fake" it responded "No", adding that they could be found on the same databases.

As fun as chatbots may be, or as advanced as they may seem, they are still prone to "hallucinations" – perfectly coherent-sounding answers that don't in any way relate to the real world.

Without heavy fact-checking, it's not really a tool you should use when trying to research a legal case that relies on real-world precedent rather than the hallucinations of a spicy autocomplete.

The lawyer wrote that he "greatly regrets having utilized generative artificial intelligence to supplement the legal research performed herein" and vows to "never do so in the future without absolute verification of its authenticity".

Both Schwartz and lawyer Peter LoDuca, who was not aware that ChatGPT had been used while researching the case, are facing a hearing on June 8 about the incident.

情報源を読む

リサーチ

  • “AIインシデント”の定義
  • “AIインシデントレスポンス”の定義
  • データベースのロードマップ
  • 関連研究
  • 全データベースのダウンロード

プロジェクトとコミュニティ

  • AIIDについて
  • コンタクトとフォロー
  • アプリと要約
  • エディタのためのガイド

インシデント

  • 全インシデントの一覧
  • フラグの立ったインシデント
  • 登録待ち一覧
  • クラスごとの表示
  • 分類法

2024 - AI Incident Database

  • 利用規約
  • プライバシーポリシー
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd