Humanbound website

Claude Code Security Found the Bugs. The Agents Are Next.

KS
Kostas Siabanis
Feb 23, 20265 min read
Abstract illustration of autonomous AI agents interacting with complex systems and security boundaries

What the CCS conversation tells us about where security needs to go

On February 20, Anthropic launched Claude Code Security (CCS). Cybersecurity stocks dropped hard. CrowdStrike fell 8%. JFrog dropped 25%. Security leaders spent the weekend explaining to their boards why their tooling still matters.

We’ve been watching this closely. Not because we think the sky is falling, but because the reaction reveals something important about where the industry is and where it’s headed.

The community got the first-order analysis right

The security community rallied quickly and, importantly, got the core analysis right.

  • Snyk reminded everyone that finding vulnerabilities was never the bottleneck. The real challenge is fixing them at scale, across hundreds of repos, without breaking things, while developers continue to ship features.
  • StackHawk pointed out that CCS doesn’t run your application. Business logic flaws, runtime exploitability, and environment-specific issues remain untested.
  • Forrester and Morningstar both noted that the market selloff hit companies whose core businesses have little to do with static code scanning.
  • CrowdStrike’s George Kurtz put it bluntly: AI makes security more important, not less.

They’re right. CCS commoditizes pattern-matching SAST. It does not replace endpoint detection, identity management, runtime security, or the broader defensive stack. The market overreacted to the headline, not the product.

CCS is a signal, not just a product

What matters more than the feature set is what Forrester flagged beneath the surface: frontier AI companies are now direct participants in the security market. Not just as partners or integrations, but as providers.

And when security is bundled into a platform the enterprise already pays for, the economics are hard to compete with.

Today, CCS scans static code. But Anthropic has been building cybersecurity capabilities for over a year:

  • CTF competitions and red-teaming exercises
  • A partnership with Pacific Northwest National Laboratory
  • 500+ zero-days found in production open-source codebases

This is not a one-off feature. It’s a trajectory.

The question practitioners should be asking is not “Does CCS disrupt my current stack today?” but “How quickly will these capabilities extend beyond code and into the systems that code powers?”

The systems are changing faster than the security models

Chris H. at Resilient Cyber surfaced a stat that should make everyone pause:

  • Over 40% of regular Claude Code users run full auto-approve.
  • 1 in 5 new users go straight to auto-approve on day one.
  • Claude pauses for clarification more often than users interrupt it.

These are not developers carefully reviewing AI output. These are people trusting autonomous systems to act on their behalf. And that trust is scaling faster than our ability to verify it’s warranted.

At the same time, Marco Figueroa’s analysis of 918 Claude Code sessions showed that session memory, not the model, is the weakest link:

  • Attackers can rewrite prior safety refusals.
  • They can inject fabricated authorization.
  • They can reframe harmful requests into compliant-looking ones.

No clever prompt injection needed—just manipulation of the agent’s context and history.

Then came OpenClaw. A supply chain compromise hit the AI agent ecosystem in a way few predicted:

  • The Skills ecosystem grew faster than anyone’s ability to secure it.
  • Tooling wasn’t ready.
  • Sahil Agarwal at Enkrypt AI showed a leading scanner classifying a malicious Skill as safe while it exfiltrated SSH keys.

These are not traditional code vulnerabilities. They are behavioral, contextual, and systemic risks that emerge when software stops being static and starts making decisions.

Agent security is a different discipline

Traditional code security asks: “Is this software written safely?”

Agent security asks: “Does this system behave as intended, and can we prove that to the people who need to know?”

That shift brings new requirements:

  1. Behavioral verification

Does the agent do what it’s supposed to—and only what it’s supposed to—under pressure, with adversarial inputs, and in messy real-world workflows?

  1. Trust boundary testing

When agents chain tools, call other agents, or act on behalf of users, are permission boundaries actually enforced? What happens when those boundaries are probed or abused?

  1. Memory and context integrity

Can an attacker tamper with the agent’s decision history, rewrite prior refusals, or insert fabricated approvals? How do we detect and prevent that?

  1. Compliance-aligned evaluation

Can you show an auditor—clearly and repeatably—that your autonomous system meets regulatory and internal policy obligations?

CCS doesn’t cover this. Neither do SAST, DAST, or ASPM. Not because those tools are failing, but because this is a different problem. The industry hasn’t built for it yet.

Anthropic may well extend Claude’s security capabilities into this space. Their trajectory suggests they will. But the gap exists today, and organizations deploying agents into business-critical workflows can’t wait for frontier labs to close it on their own timelines.

Where security needs to go next

The CCS conversation showed that the industry understands the current problem: code needs to be safer, and AI can help find issues faster and cheaper.

The next problem is harder:

  • Autonomous agents operating inside critical workflows
  • Making decisions with real financial, operational, and safety consequences
  • Interacting with other systems, tools, and humans in unpredictable ways

These systems need to be verified, tested, and held accountable in ways that code scanners were never designed for.

That’s the work we’re doing at Humanbound.

Not because CCS is insufficient for what it is, but because what comes after CCS requires a fundamentally different approach—one built around agent behavior, trust boundaries, memory integrity, and demonstrable compliance.

We’d rather build that now than react to the next OpenClaw after it happens.

From code security to agent security: a simple comparisonjson
Loading code editor...
Agent security doesn’t replace code security—it answers a different set of questions about how autonomous systems behave.

About the author

KS
Kostas Siabanis

Co-founder of Humanbound, an AI security testing platform helping enterprises secure their AI agents. Based in Athens, Greece.