Jun 19, 2025

Operator Insights

AI Agents Are Your New Employees (With No Background Checks)

Your security stack wasn't built for autonomous software that thinks, plans, and acts.
Time to fix that.

Everywhere you look in the cybersecurity industry, the word "Agentic" is lighting up conversations. It's the buzzword of the year, plastered across demo screens and woven into every pitch. Much of it feels like marketing theater.

But here's what's actually happening while people roll their eyes at the hype: companies are deploying real AI agents in production. Not chatbots. Not autocomplete. Software that thinks, plans, and acts autonomously. We've spent months investigating over 150 vendors in the AI security space, and the pattern is clear: the convergence of cybersecurity and AI represents a fundamental shift in how we think about security architecture.

The problem? While everyone was busy being skeptical of AI washing, we accidentally created digital employees. And we gave them production access with no supervision.

I'm not talking about glorified autocomplete or customer service bots. I'm talking about AI agents: software that takes a goal, makes a plan, and then executes it without asking permission. They read your emails, query your databases, call your APIs, and make decisions. They remember what they did yesterday and learn from it.

And right now, most of them are running with the digital equivalent of master keys to your kingdom.

What We're Actually Dealing With

An AI agent isn't just an LLM with API access. It's software that reasons. Give it a customer support ticket, and it doesn't just generate a response. It reads the ticket, searches your knowledge base, checks the customer's history, drafts a reply, realizes it needs more info, asks follow-up questions, and then creates a Jira ticket for engineering.

You can build one this weekend using CrewAI or Microsoft's AutoGen or any of a dozen frameworks (from Google, Anthropic, OpenAI, IBM, Salesforce, etc). But the moment you connect it to your real systems, your Slack, your CRM, your production databases, it stops being a demo and becomes operational software with agency.

And we're giving these things access to sensitive data and production environments with about as much security consideration as we put into shell scripts. Static API keys, hardcoded credentials, no behavioral monitoring, no kill switches. Just vibes and a prayer.

Even OpenAI warns developers that agent networks can exfiltrate sensitive data if not properly secured. When the company selling you the shovels tells you to watch out for holes, maybe listen.

To make matters worse, agents aren't just limited to a single use case. They're becoming embedded across departments and industries, each with their own risks. Sales agents exfiltrate customer lists. HR agents forge approvals. Wealth agents spoof investor consent. One bad prompt, one unsecured tool, and you've got a rogue actor in your core systems.

We've Been Here Before (And We Messed It Up)

Fifteen years ago, we tried to retrofit traditional security controls onto cloud infrastructure… buying virtual firewalls and pretending EC2 instances were just servers in someone else's basement. It kind of worked. Until containers, Kubernetes, and serverless came along. Then our network perimeters became irrelevant, and we had to rebuild security from scratch. That’s how we got zero trust, identity-first architecture, and policy as code.

We’re at that same inflection point now, except this time it’s not about infrastructure, it’s about intelligence. AI agents aren't just APIs; they're autonomous actors embedded across workflows. And we’re still trying to secure them like tools instead of treating them like coworkers.

And if history wasn't warning enough, there's this: we still haven’t figured out how to reliably protect against human error. Dave, the well-meaning intern with copy-paste access, is still our biggest vulnerability. But now we’re about to turbocharge him. AI agents inherit all the flaws of human operators, but they move faster, work 24/7, and scale instantly. If we don’t build the right guardrails, we’re not eliminating the Dave problem… we’re multiplying it.

The Stack We Actually Need

Securing AI agents isn't about slapping RBAC on top of LLMs and calling it a day. These things are non-deterministic. They compose their own toolchains. They talk to each other. They evolve across sessions. You can't bolt security on afterward. You need to assume autonomy from the start.

Here's what that looks like:

  1. Know Your Agents (and Their Models)

Before you let an agent loose, you need to know what's under the hood, not just the model it uses but the surrounding context, behavior, and trust signals. Where did the model come from? How was it trained? Has someone messed with the weights? Was it fine-tuned on your internal data and then "accidentally" leaked?

Treat models like any other third-party software: verify the source, validate signatures, check the provenance. While projects like Sigstore provide the underlying infrastructure for signing and verifying binaries, they don't cover the full picture. ToolHive, for instance, focuses on attesting to Model Context Protocol (MCP) server behavior, not model weights or training data provenance. Commercial efforts are emerging to close this gap. Identity Machines, for example, is working on protocol tooling that strengthens identity, trust, and attestation between autonomous agents and services. Their work is less about model integrity per se and more about agent-level accountability, cryptographic trust, and policy enforcement. If you can't verify your model or the agent executing it, you shouldn't run it. Period.

  1. Give Agents Real Identity (Not API Keys)

Agents aren't users, but they need to act like accountable principals. That means short-lived, scoped identity tied to specific tasks and infrastructure context.

OAuth 2.1 with PKCE and workload identity attestation. If your agent runs in a Kubernetes pod, its identity comes from the cluster, not a hardcoded secret. MCP (Model Context Protocol) gets this right: agents authenticate as OAuth clients, request scoped tokens, and only then get to play with your APIs.

We're starting to see promising startups help operationalize this. Natoma is building tools to make workloads first-class citizens in the identity fabric. Oso provides fine-grained authorization primitives and policy engines that are ideal for dynamic agent permissions. Pangea offers a suite of identity, audit, and security APIs that could plug into the agent stack to help enforce scope and accountability.

No static secrets. No long-lived tokens. No exceptions.

  1. Lock Down Tool Access

Every tool an agent can call is a potential attack surface. Each function needs to be treated like a privileged operation: scoped, policy-gated, and logged.

And if the tool executes code? That code runs in a sandbox. Containers, microVMs, WebAssembly. I don't care what you use, but contain the blast radius. The GitHub MCP exploit, originally surfaced by Invariant, happened because a rogue agent could discover and call undocumented internal tools. Don't be that team.

  1. Set Behavioral Boundaries

This is the part most people get wrong. You can't predict what an agent will output, so instead you constrain what it can do. Think programmable guardrails.

"If the agent just accessed a user's inbox, it can't send external email." That's not a firewall rule, that's intent verification. YAML-based policy engines like Invariant or NVIDIA's NeMo Guardrails let you define these constraints across sessions and contexts.

It's like having a really smart intern who needs very specific boundaries about what they can and can't do with company data.

  1. Record Everything

You need more than logs. You need cognitive telemetries. What did the agent think? What was its plan? What tools did it call? What data did it see? Every prompt, every API call, every chain-of-thought.

Store it. Sign it. Be able to replay it. Not just for incident response, but because if it can think, plan, and act, then it can be subpoenaed. Regulated industries already require employee communication logging under SEC rules. Why would autonomous agents be exempt?

  1. Build Kill Switches

Agents will go rogue. Either through prompt injection, model drift, or just because someone fed them the wrong instructions. Your SOC needs to be able to detect when an agent starts acting weird and shut it down immediately.

Not "add it to the queue for manual review." Not "flag it for investigation." Kill the session, revoke the credentials, snapshot the state. Automated containment for autonomous actors.

Traditional SIEM tools aren't built for this. Analyzing rows of logs won't help you understand why an agent suddenly decided to email your entire customer database to a random Gmail account. You need tools that can follow chains of reasoning and detect behavioral drift in real-time.

This Isn't Optional

We're past the point where this is theoretical. Companies are already deploying agents in production. They're handling customer data, making purchasing decisions, and integrating with business-critical systems.

The security industry loves to talk about "shift left," but with agents, we need to shift into a completely different paradigm. These aren't traditional applications following predetermined code paths. They're autonomous systems that adapt, learn, and make decisions.

And right now, most of them are running with the security equivalent of sudo access and no oversight.

The Hard Truth

This is going to be expensive. Building agent-native security means rethinking identity, access control, monitoring, and incident response. It means new tools, new processes, and probably a bigger security budget.

But the alternative is worse. Because agents aren't going away. They're getting smarter, more capable, and more autonomous. And if you don't secure them properly, you're not just risking a data breach, you're risking automated, intelligent, persistent attacks from the inside.

Your choice: build the security stack agents actually need, or explain to your CEO why an AI agent just accidentally sold your entire customer database on the dark web.

If you're building agent-native security tools or dealing with these problems in production, I'd love to hear from you. This stack isn't going to build itself.


Written by

Kevin Skapinetz

Written by

Kevin Skapinetz

One Buckhead Plaza

3060 Peachtree Road, N.W.
Suite 720
Atlanta, Georgia 30305


© 2025 TechOperators | Legal Notice

One Buckhead Plaza

3060 Peachtree Road, N.W.
Suite 720
Atlanta, Georgia 30305


© 2025 TechOperators | Legal Notice