← Back to feed
Cybersecurity

Hacker uses Claude and ChatGPT as assistant-in-the-loop to breach multiple government agencies

A threat-actor profile reported on r/technology and escalated across AI-security Twitter this weekend: an individual used Claude and ChatGPT as coding assistants to compose novel exploit chains against at least three US federal agencies. The attacker reportedly fed LLMs the target environment's architecture via open-source filings, had them generate bespoke phishing payloads and post-exploitation scripts, and iterated until bypasses worked. Anthropic and OpenAI have since rotated safety filters; Anthropic disclosed they had downgraded MCP cache TTL on March 6 specifically to shorten the window for adversarial prompt-cache poisoning. Sets the new baseline for "AI-assisted threat actor" reporting.

ClaudeChatGPTLLM AbuseFederal BreachPrompt InjectionMCP

Why it matters

This is the canonical "AI made this attack trivial" narrative that security-budget conversations have been waiting for. Even if the technical details turn out overstated, the political impact is real — expect federal guidance on LLM usage in sensitive environments within 30 days, and a new wave of enterprise policies banning personal LLM accounts on work devices. Claude and ChatGPT will face pressure to ship tighter abuse detection on code-completion and multi-step planning in the next few releases.

Impact scorecard

7.8/10
Stakes
9.5
Novelty
8.5
Authority
7.0
Coverage
7.5
Concreteness
7.0
Social
8.5
FUD risk
4.0
Coverage15 outlets · 2 tier-1
Reddit r/technology, Hacker News, Wired, Ars Technica, SecurityWeek, Dark Reading
X / Twitter4,800 mentions
@GossiTheDog · 5,100 likes
@briankrebs · 3,700 likes
Reddit4,200 upvotes
r/technology
r/technology, r/netsec, r/cybersecurity

Trust check

medium

Core claim (LLM-assisted government breach) has multi-outlet coverage and credible security-Twitter amplification — but details are still partly single-sourced and the attacker profile comes from one investigative thread. Wait for Anthropic or OpenAI's official post-mortem before treating specific capability claims as verified. Moderate FUD risk because "AI-assisted hacker" framing is politically charged.