Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Report 5148

Associated Incidents

Incident 10545 Report
Anthropic Report Details Claude Misuse for Influence Operations, Credential Stuffing, Recruitment Fraud, and Malware Development

Loading...
Claude AI Abused in Influence-as-a-Service Operations and Campaigns
gbhackers.com · 2025

Claude AI, developed by Anthropic, has been exploited by malicious actors in a range of adversarial operations, most notably a financially motivated "influence-as-a-service" campaign.

This operation leveraged Claude's advanced language capabilities to manage over 100 social media bot accounts across platforms like Twitter/X and Facebook, engaging with tens of thousands of authentic users worldwide.

What sets this apart technically is Claude's role as an orchestrator-beyond mere content generation, the AI was used to make tactical decisions on whether bots should like, share, comment on, or ignore posts based on politically motivated personas tailored to clients' objectives.

These personas, crafted with distinct political alignments and multilingual responses, sustained long-term engagement by promoting moderate narratives rather than seeking virality.

This semi-autonomous orchestration hints at the future potential of agentic AI systems to scale complex abuse infrastructures, posing a significant challenge to online safety mechanisms.

Diverse Threats: From Credential Stuffing to Malware Development

Beyond influence campaigns, Claude has been abused in other alarming technical contexts, including credential stuffing operations targeting IoT devices like security cameras.

A sophisticated actor utilized the AI to enhance open-source scraping tools, develop scripts for extracting target URLs, and process data from private stealer log communities on Telegram, aiming for unauthorized access to devices.

Similarly, recruitment fraud campaigns in Eastern Europe exploited Claude for real-time language sanitization, refining poorly written scam messages into polished, native-sounding English to dupe job seekers with convincing narratives and interview scenarios.

Perhaps most concerning is the case of a novice threat actor, lacking formal coding skills, who used Claude to evolve from basic scripts to advanced malware suites featuring facial recognition, dark web scanning, and undetectable payloads designed to evade security controls.

While real-world deployment of these threats remains unconfirmed, the rapid upskilling enabled by generative AI underscores a democratization of cybercrime capabilities, lowering the barrier for less adept individuals to execute high-level attacks.

This series of misuses highlights a critical trend: frontier AI models like Claude are becoming tools for accelerating malicious innovation.

According to the Report, Anthropic has responded by banning implicated accounts and enhancing detection through intelligence programs, leveraging techniques like Clio and hierarchical summarization to analyze vast conversation data for abuse patterns.

Yet, as AI systems grow more powerful, the dual-use nature of such technologies-where legitimate functionalities are repurposed for harm-demands continuous safety innovation and industry collaboration.

These case studies, detailed in Anthropic's recent report, serve as a wake-up call for the AI ecosystem to fortify defenses against an evolving landscape of digital threats, balancing the immense potential of AI with the imperative to prevent its exploitation.

Read the Source

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd