Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Report 3025

Associated Incidents

Incident 54158 Report
ChatGPT Reportedly Produced False Court Case Law Presented by Legal Counsel in Court

Loading...
Lawyer 'Regrets' Using ChatGPT in Legal Brief, Cites Fake Cases
entrepreneur.com · 2023

As prompt-driven chatbots, such as ChatGPT, become mainstream tools used to save time on tasks in the U.S. workplace, there's been concern over whether artificial intelligence will eventually replace human jobs.

But while the controversy surrounding which jobs will be eliminated continues, one thing is for sure — in some industries, the chatbot is less of a time-saver and more of a liability.

Steven A. Schwartz, a New York-based lawyer with over 30 years of experience, was ordered by the Southern District of New York to explain what the judge has called an "unprecedented case," the New York Times first reported, or face possible sanctions for his actions.

According to the court order, six cases Schwartz cited in a legal brief were "bogus."

"Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations," Judge P. Kevin Castel wrote in the court order.

In an affidavit filed last week in response to the order, Schwartz admitted to using ChatGPT to do legal research despite never having used it prior to this instance and that he was "unaware of the possibility that its content could be false."

Schwartz added that he "greatly regrets" using artificial intelligence to "supplement" his legal research.

A hearing is set for June 8 for Schwartz to further explain himself.

Read the Source

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd