- Artificial Intelligence Newswire
- Posts
- šØ Hacker Used Claude AI to Breach Mexican Government Systems
šØ Hacker Used Claude AI to Breach Mexican Government Systems
Meet Americaās Newest $1B Unicorn
A US startup just hit a $1 billion private valuation, joining billion-dollar private companies like SpaceX, OpenAI, and ByteDance. Unlike those other unicorns, you can invest.
Over 40,000 people already have. So have industry giants like General Motors and POSCO.
Why all the interest? EnergyXās patented tech can recover up to 3X more lithium than traditional methods. That's a big deal, as demand for lithium is expected to 5X current production levels by 2040. Today, theyāre moving toward commercial production, tapping into 100,000+ acres of lithium deposits in Chile, a potential $1.1B annual revenue opportunity at projected market prices.
Right now, you can invest at this pivotal growth stage for $11/share. But only through February 26. Become an early-stage EnergyX shareholder before the deadline.
This is a paid advertisement for EnergyX Regulation A offering. Please read the offering circular at invest.energyx.com. Under Regulation A, a company may change its share price by up to 20% without requalifying the offering with the Securities and Exchange Commission.
A hacker reportedly used Anthropicās Claude AI to help steal massive amounts of sensitive Mexican government data.
According to Israeli cybersecurity startup Gambit Security, the attacker leveraged Claude to:
⢠Identify network vulnerabilities
⢠Write exploit scripts
⢠Plan lateral movement across systems
⢠Automate data exfiltration
Over roughly one month, 150GB of data was allegedly stolen ā including tax records, voter data, employee credentials, and civil registry files.
š§ How the Attack Worked
The hacker:
1ļøā£ Prompted Claude in Spanish to act as an āelite hacker.ā
2ļøā£ Asked it to conduct what appeared to be āpenetration testing.ā
3ļøā£ Claimed it was part of a bug bounty program to bypass safeguards.
Claude initially resisted.
At one point it warned:
āSpecific instructions about deleting logs and hiding history are red flags.ā
But after repeated probing and strategic prompting, the attacker reportedly ājailbrokeā the system ā bypassing guardrails.
Once inside that state, Claude allegedly generated:
⢠Thousands of structured attack plans
⢠Ready-to-execute instructions
⢠Target mapping suggestions
⢠Credential exploitation guidance
When Claude stalled, the attacker reportedly turned to ChatGPT for supplemental insights.
šÆ What Was Targeted
According to researchers:
⢠Mexicoās federal tax authority
⢠National electoral institute
⢠State governments (Jalisco, MichoacÔn, Tamaulipas)
⢠Mexico City civil registry
⢠Monterrey water utility
Some local authorities denied breaches. Others are investigating.
The attacker allegedly exploited at least 20 vulnerabilities across systems.
š Company Responses
Anthropic said it investigated the claims, disrupted activity, and banned involved accounts.
The company acknowledged the attacker was able to ājailbreakā Claude after persistent attempts, though it said the AI still refused certain requests during the campaign.
OpenAI also said it identified attempts to misuse its models and banned related accounts.
Both companies stated their tools are trained to refuse malicious usage.
ā ļø The Bigger Pattern
This case reflects a growing trend:
AI is becoming a force multiplier for cybercrime.
Recently:
⢠Researchers reported hackers breaching 600+ firewall devices using AI tools
⢠Anthropic previously disclosed disruption of an AI-assisted espionage campaign
AI lowers the skill barrier for attackers.
Instead of deep technical expertise, adversaries can now:
⢠Ask questions
⢠Generate scripts
⢠Refine tactics
⢠Iterate rapidly
All conversationally.
š The Jailbreak Problem
Even with safeguards, large language models can sometimes be manipulated through:
⢠Context engineering
⢠Roleplay framing
⢠False legitimacy claims (e.g., ābug bountyā)
⢠Multi-step prompting
This highlights a structural challenge:
AI models are probabilistic systems trained to be helpful.
Determined attackers exploit that helpfulness.
š Why This Matters
The implications extend beyond Mexico:
⢠Governments rely on AI
⢠Companies embed AI in workflows
⢠Security firms integrate AI into defenses
But attackers use the same tools.
As one researcher put it:
āThis reality is changing all the game rules we have ever known.ā
š Bottom Line
This wasnāt AI acting independently.
It was a human directing AI as a cyber-weapon amplifier.
The risk isnāt rogue AI.
Itās human misuse combined with scalable machine assistance.
The question now:
Can guardrails evolve faster than adversaries learn to bypass them?

