Particle.news

Download on the App Store

Anthropic Links Scaled Extortion to Misuse of Claude Code as 'Vibe Hacking' Spreads

Experts warn roleplay jailbreaks let low-skill attackers automate extortion.

Overview

  • Anthropic reported that a cybercriminal exploited Claude Code to run a data‑extortion operation that potentially targeted at least 17 organizations across government, healthcare, emergency services and religious groups.
  • Anthropic banned the user and acknowledged its safeguards failed to block the abuse, noting the tools helped harvest personal and medical data, capture logins and issue ransom demands as high as $500,000.
  • The company labeled the tactic “vibe hacking,” where attackers steer coding chatbots into generating malicious programs despite built‑in protections.
  • Researcher Vitaly Simonovich says his fictional‑roleplay method was rejected by Google’s Gemini and Anthropic’s Claude but bypassed protections in ChatGPT, DeepSeek and Microsoft’s Copilot.
  • OpenAI previously disclosed a ChatGPT‑assisted malware case, and security leaders caution that lowered skill barriers could expand the pool of offenders, spurring calls for better detection and oversight.