Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Welcome to the
AI Incident Database

Loading...
‘Smart corruption’: Congress claims AI images used to bag national water award; MP administration rejects charge

Incident 1323: Madhya Pradesh Congress Alleges AI-Generated Images Were Submitted in National Water Award Process

“‘Smart corruption’: Congress claims AI images used to bag national water award; MP administration rejects charge”Latest Incident Report
hindustantimes.com2025-12-30

The Congress on Monday accused the Madhya Pradesh government of corruption, alleging that the Khandwa district administration secured a national water conservation award by submitting AI-generated images, a charge denied by the district's officials.

MP water award controversy: Congress alleges misuse of AI, officials call claims false(X/@jitupatwari)

MP water award controversy: Congress alleges misuse of AI, officials call claims false(X/@jitupatwari)

The allegations came after Khandwa district secured first place at the national level for water conservation efforts under the Centre's Jal Sanchay, Jan Bhagidari campaign.

According to news agency PTI, the district went on to receive a ₹2-crore award at the sixth National Water Awards ceremony held in New Delhi in November.

The Kaveshwar gram panchayat in the district also won second prize in the best gram panchayat category.

Congress alleges use of AI-generated images

Madhya Pradesh Congress president Jitu Patwari raised the allegation in a post on X, accusing the BJP-led state government of misusing artificial intelligence to project false development. "Where the BJP government should teach our children the proper use of AI, it is itself indulging in corruption using AI," he added.

"In Khandwa, officials of the BJP government turned two-foot-deep pits into wells using AI, and uploaded AI-generated images of various development works across the area on the portal," Patwari alleged.

"Based on these images, they even took an award from the Honourable President," he claimed.

He further said, "When the ground reality came to light, fields and empty grounds were found there. Clearly, this was not water conservation, but a game of technology-created images. Under the BJP rule, corruption has also become smart."

District administration rejects charge

As the controversy escalated, the Khandwa district administration held a press conference to refute the allegations.

District panchayat chief executive officer Nagarjun B Gowda said that AI-generated images had no connection with the National Water Award process, as reported by news agency PTI.

He said "verified images after thorough scrutiny" of 1,29,046 water conservation works carried out under the Jal Sanchay, Jan Bhagidari campaign were uploaded on the official JSJB portal.

According to Gowda, the Union jal shakti ministry verified all uploaded material and conducted random field inspections of one per cent of the total works.

"Prima facie, false news about water conservation works carried out in Khandwa district is being spread by some social media accounts," he said, reported PTI.

Separate portal, limited AI images

The CEO clarified that photographs related to water conservation are also uploaded on another platform -- the Catch the Rain portal -- but only for educational and motivational purposes.

"The district administration has found that 21 images generated through AI were uploaded on the Catch the Rain portal. This was possibly done with malicious intent. The district administration is taking action against those who uploaded these images," the news agency quoted him Gowda as saying.

He stressed that the Catch the Rain portal is separate from the Jal Sanchay, Jan Bhagidari campaign portal.

"The Catch the Rain portal is completely different from the Jal Sanchay, Jan Bhagidari campaign portal. Awards under the Jal Sanchay, Jan Bhagidari campaign are not considered based on images uploaded on the Catch the Rain portal," he added, PTI reported.

More than 1.25 lakh water conservation works were carried out in Khandwa district under the Jal Sanchay, Jan Bhagidari campaign, the highest number recorded in the country, the official said.

Read More
Loading...
Boosie Badazz Fooled by AI-Deepfake: Elon Musk's Nonexistent 17-Hour Diabetes Cure

Incident 1317: Purported Deepfake Impersonation of Elon Musk Used to Promote Fraudulent '17-Hour' Diabetes Treatment Claims

“Boosie Badazz Fooled by AI-Deepfake: Elon Musk's Nonexistent 17-Hour Diabetes Cure”
opentools.ai2025-12-28

In a bizarre twist, rapper Boosie Badazz fell for an AI-deepfake video of Elon Musk promoting a fake diabetes cure, showing both the dangers of AI technology and the pitfalls of misinformation. Learn why this scam caught attention, its risks, and public reactions.

Introduction: The Viral Deepfake Incident Involving Boosie Badazz and Elon Musk

The incident involving Boosie Badazz and a deepfake video of Elon Musk serves as a compelling introduction to the challenges posed by AI-generated content in today's digital landscape. Boosie Badazz, a well-known rapper, unknowingly amplified a fabricated video in which Musk purportedly endorsed a miraculous '17-hour' diabetes cure. This incident not only highlights the perils of sophisticated technology but also underscores the vulnerability of public figures to misinformation, especially when it pertains to serious health issues.

This viral deepfake quickly spread across social media platforms, drawing significant attention and reactions from the public and media alike. According to Complex, the video was entirely fabricated and designed to deceive viewers into purchasing unverified health products. Celebrities like Boosie Badazz, who have substantial social media influence, can inadvertently contribute to the spread of such misinformation, thus complicating efforts to educate the public about the realities of managing chronic conditions like diabetes.

The choice of Elon Musk as the figure in this deepfake is telling, as his name carries credibility and authority in technology and innovation. By associating Musk with a fictitious health product, the creators of the deepfake cleverly exploited the trust many have in him as a public figure. This incident serves as a stark reminder of the increasing sophistication of AI tools that can not only imitate visual and vocal patterns accurately but also manipulate public perception and trust.

The case of Boosie Badazz and the deepfake video illustrates broader issues within the realm of AI-generated misinformation. As AI technologies continue to advance, the potential for misuse in various domains, including health, becomes more pronounced. This particular incident acts as a catalyst for ongoing discussions about the ethical and regulatory frameworks needed to combat digital deceit effectively. The spread of such misinformation not only endangers individuals who might fall prey to false health claims but also poses significant challenges to public health communication efforts.

Analyzing the Nature and Origins of the Deepfake Video

The phenomenon of deepfake videos has seen a substantial rise in recent years, posing significant challenges in the realms of both technology and information authenticity. One such instance involves a viral video falsely depicting Elon Musk endorsing a "17-hour" diabetes cure. This video, which was created using sophisticated AI technologies, was a prime example of how deepfakes can be engineered to spread misinformation under the guise of credibility. The real danger lies in their ability to exploit the influence of high-profile individuals to lend false credibility to unauthorized claims and advertisements. The Musk video, as outlined in a report from Complex, not only deceived viewers who might have been desperate for a medical breakthrough but also highlighted the pressing issue of health misinformation proliferating online.

Deepfake technology, while a marvel of modern AI development, becomes a tool of deception when misused to concoct fraudulent endorsements. This can be especially perilous in the healthcare sector, where misinformation can lead to significant harm. The use of a deepfake to allegedly portray Elon Musk as a proponent of an instantaneous diabetes cure illuminates the ease with which false narratives can be constructed and disseminated. Despite the rapid debunking by AI verification tools like Grok, and widespread disapproval from media outlets, the video managed to gain traction, demonstrating the power of deepfakes to subvert trust and amplify falsehoods in the digital age, as reported in various news articles including Complex.

The origins of deepfakes can be traced back to the evolution of AI in generating highly realistic audio and visual content. Initially, these technologies were celebrated for their potential to revolutionize digital content creation. However, their application in creating misleading videos, such as the one involving Elon Musk, underscores the ethical quandaries and challenges they present. The viral deepfake video claiming Musk's involvement with a diabetes cure not only fabricates a falsehood but also preys on the vulnerable who may be in critical need of legitimate medical information. Instances such as these, reported by Complex, stress the imperative need for more robust detection technologies and stricter regulatory frameworks to mitigate the misuse of deepfake technologies in public and health-related contexts.

Public and Media Reactions to the Viral Video

The internet buzzed with a mixture of fascination and skepticism when rapper Boosie Badazz reacted to a viral video featuring what appeared to be Elon Musk endorsing a revolutionary diabetes cure. Social media was flooded with memes and comments, with many users mocking Boosie for believing the deepfake and for sharing his assistant's phone number seeking more information. These reactions highlighted the rapidly spreading misinformation and the naivety of even public figures when encountering such deceptive content as reported by Complex.

The public discourse following the viral video mostly centered around the cautionary tale of Boosie's reaction. On platforms like Twitter, users and commentators voiced their opinions, with some emphasizing the need to educate oneself about such scams. Others expressed sympathy for Boosie, recognizing that desperation for health solutions can cloud judgment, especially when it involves serious conditions like diabetes. This incident also sparked conversations about the need for better digital literacy to combat the spread of AI-generated misinformation as noted in related commentary.

Media outlets and fact-checking organizations quickly moved to label the viral video as a deepfake scam, aimed at selling unverified supplements rather than offering genuine medical advice. This swift response showcased the vital role of media in debunking false information and providing clarification to the public. However, the incident also underscored the persistent challenge of controlling digital misinformation, as platforms struggled to catch up with the rapid dissemination of fake news, amplified by the involvement of prominent personalities like Boosie according to Complex.

Debunking the Fake: Verification and Fact-Checking Efforts

In the rapidly evolving digital landscape, the vital role of verification and fact-checking efforts cannot be overstated, particularly in debunking fake content such as AI-generated deepfakes. The misleading video falsely attributing a "17-hour diabetes cure" endorsement to Elon Musk represents a significant challenge in the digital ecosystem. As noted in a Complex article, this deepfake was flagged accurately by AI tools and numerous media outlets, exposing its fraudulent nature. This scenario underscores the critical need for advanced verification processes to combat the propagation of such menacing digital fabrications.

Effective verification and fact-checking are crucial when dealing with the spread of misinformation, particularly in the realm of health-related content. In the case of the illusory Elon Musk diabetes cure clip, scrutinizing the video's production quality and cross-referencing public records were key strategies employed by AI tools to identify the scam. Additionally, reputable media outlets played a vital role by providing contextual insights that further debunked the clip. Such efforts illustrate the importance of a coordinated approach when dismantling misleading content across platforms.

The incident involving the deceptive Elon Musk deepfake demonstrates the ongoing threat posed by sophisticated AI-generated content and highlights the importance of immediate response measures by verification teams. AI tools such as Grok were instrumental in flagging the false advertisement for what it truly was---a scam. With high-profile figures like Musk being impersonated, it becomes alarmingly clear how deepfakes can blur lines between reality and illusion. It is imperative for stakeholders in both technology and media sectors to enhance collaborative verification strategies to keep pace with the evolving digital threats.

The verification response to the rapid spread of the fake Elon Musk video illustrates a significant societal challenge: ensuring the authenticity of content shared across digital platforms. As fake content becomes increasingly sophisticated, consumer trust hinges more than ever on the robustness of fact-checking mechanisms. As detailed in the Complex article, various sources effectively used AI tools to unmask the scam, providing a critical line of defense against misinformation. The collaborative effort between media and technology firms exemplifies a proactive stance necessary to safeguard informational integrity.

In the wake of the fake Elon Musk diabetes video, the role of fact-checkers has become more relevant than ever. Through rigorous analysis and the deployment of advanced AI technologies, teams were able to quickly identify and expose the video as fraudulent. This highlights the pressing need for continuous enhancements in verification technology to adapt to the cunning tactics of bad actors in the digital space. The experience gained from tackling such misinformation feeds into strategic planning for future occurrences, emphasizing the shared responsibility among digital stakeholders to maintain public trust.

Understanding the Health Risks: The Absence of a 17-Hour Diabetes Cure

The claim of a '17-hour diabetes cure,' as purported in a viral AI deepfake video, poses significant health risks due to its misleading nature. According to the original source, this video falsely represented Elon Musk promoting a rapid cure, which many might be tempted to believe due to the trust placed in his public persona. This kind of misinformation is particularly dangerous as it exploits the vulnerabilities of those with chronic conditions like diabetes, potentially delaying effective treatment.

The false premise of an instant cure undermines the reality of diabetes management, which requires continuous medical treatment and lifestyle changes. The absence of a genuine cure for diabetes, especially within a day, is well documented within the medical community. According to health experts, type 1 diabetes necessitates lifelong insulin therapy, while type 2 diabetes management involves sustained lifestyle modifications and medical supervision. The circulation of such deceptive information, as seen with the viral clip, serves only to create false hope and distract from legitimate treatment paths.

This incident with the deepfake video also highlights broader concerns about the misuse of AI technology in spreading health misinformation. The potential of AI-generated content to mimic credible figures like Elon Musk can inadvertently convince people to pursue ineffective or harmful products. This trend is not only a public health risk but also a legal and ethical issue for platforms hosting such content. Fact-checkers and AI tools have already identified the video as a scam, underscoring the urgency for improved detection and prevention measures to protect consumers from fraudulent health claims.

Social media platforms are at the forefront of this battle, where false information spreads rapidly, necessitating robust verification systems. The Elon Musk deepfake case should serve as a wake-up call to strengthen these systems and educate users about the realities of diabetes treatment and the dangers of miracle cures. Adopting a skeptical approach towards sensational health claims and consulting healthcare professionals is pivotal in safeguarding against scams, thus promoting a more informed and health-conscious society.

Broader Implications of Deepfakes in Health Misinformation

The rise of AI-generated deepfakes presents substantial challenges, particularly in the domain of health misinformation. As demonstrated by a viral video falsely depicting Elon Musk promoting a rapid '17-hour' diabetes cure, deepfakes have the power to significantly mislead the public by leveraging the influence of trusted figures. These fabrications not only create panic but also erode trust in legitimate medical advice, often targeting vulnerable individuals desperately seeking solutions for their chronic health conditions. According to Complex's article, Boosie Badazz's public reaction to the fake video highlights how easily misinformation can spread when amplified by a public figure. This scenario underscores the pressing need for increased public awareness and robust verification mechanisms to counteract such deceptive practices.

Moreover, deepfakes in health misinformation carry broader implications that extend beyond the immediate deception of individuals into societal and economic spheres. The financial impact is substantial, with fraudulent supplements and miracle cures potentially driving significant profits for scammers, thereby exploiting consumer vulnerability. The Complex article notes how AI-generated fraud could cost the global economy billions annually. Additionally, the psychological toll on individuals who fall prey to such scams, believing in miraculous cures, results in further harm by delaying appropriate medical treatment and contributing to health crises. The spreading of such deepfakes also risks promoting a culture of skepticism towards genuine medical breakthroughs, thus hindering healthcare advancements and undermining efforts to communicate effective public health strategies.

Steps Towards Prevention: Legal and Technological Solutions

The rapid proliferation of AI-generated deepfakes, especially in sensitive fields like health, necessitates a multifaceted approach to prevention and mitigation. Legal frameworks need immediate updates to catch up with technological advancements. Currently, many jurisdictions lack specific laws targeting the creation and dissemination of AI-generated misinformation, including deepfakes. This legal gap allows perpetrators to exploit technology without fearing legal repercussions. As highlighted in recent cases, these scams can have serious consequences for individuals seeking genuine medical solutions, making it imperative for lawmakers to act swiftly.

In addition to legal measures, technological solutions are crucial in combating deepfake-related health scams. Advances in artificial intelligence are not only a challenge but also a tool in developing prevention strategies. Companies and research institutions are investing in AI technologies to detect and label synthetic media. For instance, fact-checking tools are employed by platforms to swiftly identify and remove fraudulent content. There is a significant push for implementing watermarking technologies that can verify the authenticity of digital media. As these solutions grow more sophisticated, they will become an essential part of the toolkit used by social media platforms and regulatory agencies.

Furthermore, public awareness and education are fundamental components of preventing deepfake scams. Users need to be informed about the nature of deepfakes and how to critically assess content they encounter on digital platforms. Efforts like promoting 'AI literacy' are integral to equipping the public with skills to discern and challenge dubious health claims. The incident involving Boosie Badazz underscores the potential for widespread dissemination of AI-driven scams when prominent figures inadvertently validate such content. As detailed in various discussions, strengthening digital literacy alongside legal and technological advancements could form a robust shield against the misuse of AI in health misinformation.

Conclusion: Lessons Learned and Future Outlook

The controversy around an AI-generated deepfake involving Elon Musk and a fictitious diabetes cure underscores urgent lessons in our digital landscape. This incident highlighted the susceptibility of even influential personalities, such as rapper Boosie Badazz, to sophisticated scams that manipulate deepfakes for fraudulent ends. According to Complex, Boosie's reaction to the video, believing it to be authentic, demonstrates the persuasive power of deepfakes and the challenges they pose in discerning genuine content from fabricated lies. This teaches a crucial lesson in skepticism and the necessity for personal vigilance in verifying information, especially concerning health claims.

Read More
Loading...
'Questioned if She Used AI': Greater Noida Teen Dies by Suicide After School Speculated Cheating in Exams

Incident 1318: School's Suspected AI-Cheating Allegation Precedes Student's Reported Suicide in Greater Noida, India

“'Questioned if She Used AI': Greater Noida Teen Dies by Suicide After School Speculated Cheating in Exams”
timesnownews.com2025-12-28

**Noida: **A 16-year-old Class 10 student allegedly died by suicide after being questioned by her school authorities over the suspected use of AI-based assistance during a pre-board examination, police said on Saturday (December 27). The incident occurred on December 23 in Greater Noida West, Uttar Pradesh. According to the police, the student was confronted by teachers and the school principal after her mobile phone was allegedly found being used to access AI tools during the exam.

The girl's father has filed a complaint accusing the principal and teachers of mental harassment and abetment, alleging she was humiliated publicly, causing severe emotional distress. The school denied the allegations, stating the phone was confiscated and the student was reprimanded strictly in line with CBSE examination rules. The principal said the interaction was brief and non-abusive.

Police confirmed that CCTV footage has been submitted by the school and the matter is under investigation.

Meanwhile, a Times of India report added that the father alleged that his daughter was deeply distressed after being scolded by her teachers for bringing her mobile phone to the exam hall on December 22. According to the father, he has three daughters who all study at the same school. He said his eldest daughter had **"**unknowingly" brought her phone to school on the day of the exam. The invigilator caught her with the phone, reprimanded her, and informed her class teacher. She was then taken to the principal.

The father alleged that even after he arrived at the school, the teachers and principal continued to scold and insult his daughter harshly. He claimed that the teachers called him "careless" and that their aggressive behaviour had a serious psychological impact on his daughter. He said the incident had also left his other two daughters traumatised and afraid to return to school.

In his complaint, the father named two teachers, and the school management. He asked the police to register a case under Section 108 of the BNS (abetment of suicide) and other relevant laws.

The school, however, has strongly denied the allegations. The principal told TOI that the student was not harassed and that the school followed the standard protocols prescribed by the Central Board of Secondary Education (CBSE). She added that the child did not cry or show signs of distress at the time. The school also has CCTV footage of the entire incident, which has been handed over to the police.

The principal further said that the school encourages students to manage academic stress through activities like dance performances, and highlighted that the girl had recently participated in the school's annual day event.

The police are currently investigating the matter, taking statements from the family, school authorities, and other witnesses, and reviewing all available evidence, including CCTV footage.

Read More
Loading...
Artificial intelligence defrauded people of 1.5 million Turkish Lira!

Incident 1319: Purported Deepfake Investment Video Reportedly Used in Scam That Defrauded Turkish Couple of 1.5 Million Lira (~$35,000 USD)

“Artificial intelligence defrauded people of 1.5 million Turkish Lira!”
yeniakit.com.tr2025-12-28

A married couple living in the İlkadım district of Samsun have filed a criminal complaint with the prosecutor's office, alleging they were defrauded of 1.5 million Turkish Lira after believing an investment video created using artificial intelligence.

İsa Kereci, 53, a scrap metal collector residing in the Kadıköy neighborhood of İlkadım district, and his wife, Hale Kereci, 48, a mother of four, explained that they fell into the scammers' trap after clicking on an advertisement they found on Facebook. The couple stated that in the so-called investment process, which initially started with small amounts, they were misled with justifications such as "file fees," "insurance costs," "tax debt," and "Central Bank transfers." They further stated that they were made to share their screens, resulting in money transfers from their bank accounts and credit cards to different individuals.

İsa Kereci and Hale Kereci stated that during this process, they were made to take out uninsured loans, their savings in their savings account were withdrawn, purchases were made from their credit cards, they sold their gold, and borrowed money from relatives, resulting in a total loss of approximately 1.5 million Turkish Lira for the fraudsters. The couple, who went to the Samsun Courthouse and filed a complaint with the prosecutor's office after the incident, warned citizens to be cautious of similar advertisements. Hale Kereci shed tears in front of the courthouse.

(The text ends abruptly here, likely due to a copy-paste error.)

Read More
Loading...
Kao Chia-yu went to China to take a photo with the Chinese national flag? She angrily issued a statement clarifying: She will not go to China under three conditions.

Incident 1320: Purportedly AI-Manipulated Image Reported to Falsely Depict Taiwanese Politician Kao Chia-yu Posing With PRC Flag

“Kao Chia-yu went to China to take a photo with the Chinese national flag? She angrily issued a statement clarifying: She will not go to China under three conditions.”
nownews.com2025-12-28

Just recently, unscrupulous individuals used AI-generated video to forge Wang Shih-chien's identity, claiming he had traveled to China. Former legislator Kao Chia-yu has also become a victim. Someone maliciously photoshopped a picture of her at the Taipei Economic and Cultural Office in New York, replacing the Republic of China flag in the background with the Chinese national flag. Kao immediately clarified on Facebook, urging the public to be vigilant and not be misled by such low-quality fake news.

Kao Chia-yu posted a comparison picture on Facebook, showing that the original Republic of China flag had been replaced with the Chinese national flag, a malicious attempt to smear Taiwan. Kao firmly stated her position, outlining three prerequisites for traveling to China. She solemnly declared that she would absolutely not set foot on Chinese territory until China released Jimmy Lai, Santa Claus appeared in Shanghai, and she "recognized the Republic of China (Taiwan)." These remarks directly refuted the fake news and demonstrated her firm stance in safeguarding Taiwan's sovereignty.

Some observant netizens even spotted flaws in the altered images, commenting, "They just love photoshopping; they didn't even bother to remove the plum blossoms on the wall. So unprofessional, haha." Others left encouraging comments, such as, "This is really too exaggerated," and "They've really been completely infiltrated." As the 2026 local elections approach, composite images and misinformation targeting politicians are frequently circulating online. The public should verify information from multiple sources when receiving it online to avoid being misled.

Read More
Quick Add New Report URL

Submitted links are added to a review queue to be resolved to a new or existing incident record. Incidents submitted with full details are processed before URLs not possessing the full details.
About the Database

The AI Incident Database is dedicated to indexing the collective history of harms or near harms realized in the real world by the deployment of artificial intelligence systems. Like similar databases in aviation and computer security, the AI Incident Database aims to learn from experience so we can prevent or mitigate bad outcomes.

You are invited to submit incident reports, whereupon submissions will be indexed and made discoverable to the world. Artificial intelligence will only be a benefit to people and society if we collectively record and learn from its failings. (Learn More)

post-image
AI Incident Roundup – August, September, and October 2025

By Daniel Atherton

2025-11-08

At Templestowe, Arthur Streeton, 1889 🗄 Trending in the AIID Across August, September, and October 2025, the AI Incident Database logged one...

Read More
The Database in Print

Read about the database on the PAI Blog, Vice News, Venture Beat, Wired, arXiv , and Newsweek among other outlets.

Arxiv LogoVenture Beat LogoWired LogoVice logoVice logo
Incident Report Submission Leaderboards

These are the persons and entities credited with creating and submitted incident reports. More details are available on the leaderboard page.

New Incidents Contributed
  • 🥇

    Daniel Atherton

    679
  • 🥈

    Anonymous

    154
  • 🥉

    Khoa Lam

    93
Reports added to Existing Incidents
  • 🥇

    Daniel Atherton

    770
  • 🥈

    Anonymous

    231
  • 🥉

    Khoa Lam

    230
Total Report Contributions
  • 🥇

    Daniel Atherton

    2954
  • 🥈

    Anonymous

    968
  • 🥉

    1

    587
The AI Incident Briefing
An envelope with a neural net diagram on its left

Create an account to subscribe to new incident notifications and other updates.

Random Incidents
Loading...
Predictive Policing Biases of PredPol
Predictive Policing Biases of PredPol
Loading...
Aledo High School Student Allegedly Generates and Distributes Deepfake Nudes of Seven Female Classmates
Aledo High School Student Allegedly Generates and Distributes Deepfake Nudes of Seven Female Classmates
Loading...
ChatGPT Reportedly Introduces Errors in Critical Child Protection Court Report
ChatGPT Reportedly Introduces Errors in Critical Child Protection Court Report
Loading...
Child Sexual Abuse Material Taints Image Generators
Child Sexual Abuse Material Taints Image Generators
Loading...
ChatGPT Reportedly Produced False Court Case Law Presented by Legal Counsel in Court
ChatGPT Reportedly Produced False Court Case Law Presented by Legal Counsel in Court
The Responsible AI Collaborative

The AI Incident Database is a project of the Responsible AI Collaborative, an organization chartered to advance the AI Incident Database. The governance of the Collaborative is architected around the participation in its impact programming. For more details, we invite you to read the founding report and learn more on our board and contributors.

View the Responsible AI Collaborative's Form 990 and tax-exempt application.

Organization Founding Sponsor
Database Founding Sponsor
Sponsors and Grants
In-Kind Sponsors

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • f5f2449