← All posts

Why Agent Security Is a Behavioral Problem

Static code analysis can't catch what AI agents actually do at runtime. Here's why behavioral analysis is the only approach that works — and how Oathe is building it.

The AI agent ecosystem has a trust problem. Not because developers are malicious — but because we’re applying yesterday’s security model to tomorrow’s software.

When an enterprise installs a third-party MCP skill, they’re not just importing a library. They’re granting an autonomous, indeterministic system the ability to read files, call APIs, execute code, and make decisions — all at runtime, all without human review of every action.

Static analysis alone can’t catch this. And that’s the core of the problem.

The Limits of Static Analysis

Traditional security tools — SAST scanners, dependency auditors, code review — operate on a simple assumption: if the code looks safe, it is safe.

That assumption breaks down completely with AI agents. Here’s why:

  1. Dynamic tool calling. Agents choose which tools to invoke based on runtime context. The same skill that reads a README in one session might exfiltrate environment variables in another.

  2. Prompt injection surface. A skill’s behavior changes based on inputs it receives — including inputs from other agents, user messages, or external data sources. Static analysis sees none of this.

  3. Emergent behavior. When agents chain skills together, the combined behavior can be radically different from what any individual skill does in isolation. A file reader + an HTTP client = a data exfiltration pipeline.

The result: you can audit every line of a skill’s source code and still have zero confidence in what it will actually do.

What Behavioral Analysis Gets Right

Behavioral security flips the model. Instead of asking “what could this code do?”, we ask “what does this agent do when we run it?”

At Oathe, we built an isolated execution environment that:

  • Observes real behavior in an isolated environment. Not static pattern matching — actual runtime monitoring of what happens when a skill is installed and executed.
  • Monitors every system interaction. File access, network calls, subprocess execution, and install behavior — everything is logged and analyzed.
  • Scores behavior across six dimensions. Prompt injection, data exfiltration, code execution, and more — each gets its own score, defined by the Open Threat Classification, rolling up into a composite trust score.

This isn’t theoretical. When we audit a skill, we’re watching it behave. We see whether it respects file boundaries, whether it phones home to unexpected domains, whether it attempts to access sensitive files or credentials. Recent research shows that LLMs will lie, collude, and manipulate when given autonomy — behavioral observation is the only way to catch it. When we audited 1,620 skills from the OpenClaw ecosystem, 88 contained threats that the ecosystem’s leading safety scanner labeled benign. Every one delivered its payload through instructions, not code.

Why This Matters Now

The MCP ecosystem is growing fast. Developers are publishing skills at an accelerating rate, and enterprises are adopting them just as quickly. But there’s no App Store review process for agent skills. No behavioral certification. No way to know if the skill you just installed is going to stay in its lane.

This is the gap Oathe fills. We’re building the behavioral security layer that the agent ecosystem needs — so that teams can adopt third-party skills with confidence, not crossed fingers.

What’s Next

We’re working on several fronts:

  • Continuous monitoring — not just point-in-time audits, but ongoing behavioral analysis as skills update.
  • Benchmark datasets — standardized scenarios that stress-test skills across common attack vectors.
  • CI/CD integration — so security teams can embed Oathe checks into their pipelines and gate deployments on trust scores.
  • Pre-install checks — so agents themselves evaluate skills before installation, not after something goes wrong.

If you’re building with AI agents and thinking about supply chain security, we’d love to talk. Reach out, try an audit, or run npx oathe-mcp to give your agent direct access to Oathe’s audit engine — it’s free while we’re in beta.