The Agent Virus - 10x coding output is also a 10x attack surface

Social engineering used to target people. Now it targets AI agents, and machines don't get suspicious.

That's the unspoken part of the vibe-coding revolution. Every conversation about AI-generated code focuses on productivity gains. Almost none focus on the attack surface that speed creates.

Andrea Michi , co-founder and CTO of depthfirst – a company that finds vulnerabilities the way elite white-hat hackers do – has been watching this space. They even found a critical one-click vulnerability in OpenClaw , the most-popular AI personal assistant (details). 

The misconception

Most assume AI-generated code carries the same risk profile as human-written code. It doesn't. Not because models write worse code (they might, they might not), but because the volume and review dynamics are fundamentally different.

When engineers wrote code manually, every line got read. With coding agents like Codex and Claude Code, teams parallelize ten agents simultaneously and ship. More code, less review, even no review!

As Andrea puts it:

"The quality of code reviews is going dramatically down."

What were seeing

The attack surface is compositional, not modular. Checking code security in isolation misses the point. Security flaws live in how services connect. One component assumes another handles authentication. A third assumes sanitized inputs. Each is correct in isolation. Together, they're a breach.

"You can write a piece of code that is completely correct... there is a vulnerability in the combination. The plumbing is where vulnerabilities are."

This is why multiple "low severity" findings, each dismissed individually, chain into exploits. It requires reasoning across components, not within them.

Agents themselves are now attack targets. More software is being driven by AI agents that interpret inputs, make decisions, and call APIs. These agents can be manipulated through prompt injection – carefully crafted inputs that trick systems into revealing data or taking unauthorized actions. The attack surface now includes machine gullibility, at scale.

Elite hacking is being democratized. iPhone-level exploitation (the kind that historically required specialists hired by nation-states) is being commoditized by AI trained on historical vulnerabilities. depthfirst is building exactly the elite defense capability: learning from synthetic vulnerable environments, then applying it to real codebases to make them secure.

But the same capability works in reverse. Vulnerabilities that were safe because no one took time to find them are now threats, because someone, anyone, can run AI to spot them.

The asymmetry

Offense and defense are both accelerating, but at different rates. Code generation is fast, frictionless, and getting cheaper. Security review is slow, contextual, and expensive. Every vibe-coded service adds surface area. Every unreviewed commit compounds the debt.

Andrea's take:

"Every software is inherently vulnerable."

The question isn't whether your code has flaws. It's whether you find them before someone else does.

What to do instead

Match your security capacity to your code generation rate. If your team's output tripled this year, did your security review capacity triple too? If not, you've systematically expanded your exposure. Tools like DeepMind's CodeMender, which automatically detects and patches vulnerabilities, signal where this is heading (details)

From scanning to compositional reasoning. Most security tooling checks code in isolation. The real vulnerabilities are in integration. You need visibility into how services interact and what each assumes about the others. This is where AI-powered security earns its place.

Treat agent security as a first-class concern. If you're deploying AI agents that make decisions or call APIs, those agents need the same adversarial testing you'd give a public-facing application. Prompt injection is already happening — most organizations just aren't looking for it yet.

Takeaway

Before celebrating your team's 10x output from AI coding tools, ask: “Has your security capacity scaled at the same rate as your code generation?

If the answer is no, you haven't built a productivity engine. You've built a larger target.

We are seeing this play out in practice across our customer base. Over the past two months, the share of new code commits containing (verified / accepted) high or critical vulnerabilities has risen meaningfully as more and more companies are using co-pilots and coding agents to write software. Many of these issues look benign but are severe in the context of the organization. Several recent examples depthfirst found looked totally benign in isolation and would have been missed by most other tools including AI tools that rely on models like Opus 4.6. But given the company’s specific AWS configuration, the interaction turned it into a very severe vulnerability.

To view or add a comment, sign in

More articles by Jack Nikodem

  • Who buys help and who doesn't

    Founders built their companies by figuring things out. The same resilience that gets a company from zero to Series B…

    2 Comments
  • FujiSoft: The $4B bet on Japan's digital drag

    KKR paid $4B for a Japanese System Integrator (SI) that is built to resist change. The headline multiple looks…

    1 Comment
  • Benesse: the neo-edtech bet

    EQT Group bought Benesse Corporation in 2023 for almost $2B – Japan's largest education company and one of Japan's…

  • Yayoi: The accounting playbook, Japan edition

    KKR has run the accounting software incumbent playbook before – MYOB in Australia, Visma in Scandinavia, Exact in the…

    1 Comment
  • The filter upstream changed. You'll feel it in 18 months.

    Seed investors are the intake valve of the venture stack. What they select today is what shows up in your growth…

  • Follow the incentives, not the pitch deck

    VCs have access to non-public financials, board-level strategy, and real-time metrics across dozens of companies. They…

  • The requirements problem

    Every major AI lab now hires forward deployed engineers. Most of them are optimizing for implementation speed when the…

    1 Comment
  • Vibe-coding changed. Leaders didn't notice.

    A lab scientist at a biotech needed something simple: a three-class image classifier behind an API. An engineering…

    3 Comments
  • Figma is being eaten

    Not by another design tool — by designers who code. Nick Budden, a design-oriented entrepreneur who’s built multiple…

  • Designing life, one day at a time

    It started 2 years ago with a single push-up and built on gradually and painfully. My inspiration came from the…

Others also viewed

Explore content categories