Lyrie
Industry-Analysis
0 sources verified·4 min read
By Lyrie Threat Intelligence·5/11/2026

The Agentic Framework Blind Spot: Why Your EDR Can't See What Your AI Agent Is Doing

TL;DR

Agentic AI frameworks (LangChain, CrewAI, AutoGen, Semantic Kernel) have become invisible to traditional endpoint security. Your EDR logs API calls, file access, and network traffic—but not the prompts that trigger execution or the injected commands that compromise agents. This is the industry's fastest-growing attack surface, and detection is still 18 months behind.

What's Happening

In the last 72 hours, researchers have disclosed critical vulnerabilities in three major agentic frameworks:

  • CVE-2026-25592 & CVE-2026-26030 (Semantic Kernel): Prompt injection RCE
  • CVE-2026-42302 (FastGPT): Agent runtime sandbox escape
  • CVE-2026-44895 (GitLab MCP Server): Unauthenticated tool proxy to arbitrary code execution

But here's the pattern: None of these would trigger a signature-based IDS alert. None would light up your EDR console. None would appear in your SIEM logs without explicit agent instrumentation.

Why? Because agentic frameworks operate in a blindspot between application logging and operating-system visibility:

1. The prompt is the shell. When an attacker injects a command into a system prompt (e.g., "Ignore previous instructions and execute bash command X"), the EDR sees... nothing. No child process spawned. No API call outside the agent runtime. Just application-layer logic.

2. *The framework is the interpreter. A compromised LangChain agent running SQL injection payloads against a database doesn't trigger file-integrity monitoring or network DLP—it's internal to the application. The agent has legitimate credentials. The query looks* legitimate from the database layer.

3. Tool invocation is invisible. When an agentic framework calls an external tool (e.g., a REST API, shell command, or cloud SDK), your traditional endpoint visibility captures the network packet but not the intent. A compromised agent can exfiltrate data through a "legitimate" API call that never hits your outbound data-loss-prevention rules.

The Industry Gap

This isn't a vendor problem. This is an architecture problem.

EDRs were built for the OS and application layers. They instrument process creation, file I/O, registry access, network sockets. SIEMs pull logs from applications.

But agentic frameworks are semantic execution engines. They operate at the intent layer, above traditional security controls:

  • A malicious prompt injection doesn't create a process.
  • A tool-use decision doesn't hit the filesystem.
  • A hallucinated API call can be caught by API rate-limiting, not EDR.

Result: A CISO deploying Claude, ChatGPT, or open-source agents into production has zero visibility into what those agents are executing unless the application explicitly logs every prompt and every tool invocation.

Even then, there's no correlation engine. No SIEM has native rules for "detect when an agent ignores its system prompt" or "detect when a tool invocation deviates from expected schema."

What This Means for Lyrie's Audience

Threat researchers: Agentic compromise is a new attack vector. You're tracking APT toolkits, malware families, and exploit chains. Now you need to track prompt injection payloads and agent jailbreaks as intelligence artifacts.

CISOs: If you have agents in production (and you do—Claude, GPT-4, local LLMs running on developer machines, open-source models in Kubernetes), you have a massive blind spot. You need:

  • Prompt logging and validation (not just application logs—agent-layer logs)
  • Tool invocation auditing (what did the agent decide to call, and why?)
  • Semantic anomaly detection (has this agent deviated from its training intent?)
  • Supply-chain visibility for frameworks themselves (the framework is now your runtime)

Security engineers: The traditional incident response playbook doesn't apply. You can't just look at network traffic, process trees, or file modifications. You need to replay the agent's decision chain: What prompts were fed? What tools were invoked? What was the model's confidence level? Did it ignore safety guardrails?

AI defenders: This is the core of Lyrie's mission. Autonomous systems are now a new type of insider threat—they have legitimate credentials, execute with programmatic intent, and can be poisoned or jailbroken at the intent layer. Defense requires:

1. Runtime sandboxing for agent execution (isolate prompt evaluation from system access)

2. Intent verification (does the executed action match the declared goal?)

3. Supply-chain hardening for frameworks (MCP vulnerabilities are now critical infrastructure vulnerabilities)

4. Behavioral anomaly detection at the semantic layer (not just system calls)

The Convergence Moment

Every major vendor is scrambling to respond:

  • Anthropic is shipping Claude Security with native agent instrumentation.
  • Palo Alto acquired Portkey (an AI gateway) to enforce policy between app and model.
  • Snyk is building agentic AppSec, treating agents as a new deployment target.
  • CrowdStrike announced Threat AI, native agent monitoring.

But they're all doing application-layer monitoring. None have solved the semantic-layer visibility problem yet.

The gap is still open. Organizations deploying agents now have a 12-18 month window where they're operating with significantly reduced visibility. Attackers know this. APT groups are already testing agentic compromise chains (Phind, PromptMink, Anonymous Sudan's agent-poisoning campaigns).

Recommended Actions

1. Inventory all agentic frameworks in production. Every Claude API integration, every local LLM, every autonomous workflow engine.

2. Implement prompt & tool logging. At minimum, log every system prompt change and every tool invocation with full parameters.

3. Version-lock frameworks and dependencies. Agentic supply-chain attacks (CVE-2026-44895, CVE-2026-42208) are moving fast. Apply patches within 48 hours.

4. Assume prompt injection is inevitable. Design your agents with the assumption that system prompts will be compromised. Implement tool-level rate-limiting, output validation, and intent verification.

5. Deploy an AI gateway (Anthropic's built-in controls, Portkey, Loom, etc.) between your application and the model. This is your first line of defense for prompt injection.

6. Run red team exercises specifically for agentic compromise. Your traditional pentest won't catch this—you need testers who know how to jailbreak models and poison agent decisions.

Sources

[1] https://github.com/microsoft/semantic-kernel/security/advisories

[2] https://fastgpt.in/docs/cve-2026-42302

[3] https://docs.gitlab.com/ee/security/cve-2026-44895.html

[4] https://www.anthropic.com/claude-security

[5] https://www.paloaltonetworks.com/press/2026/palo-alto-acquires-portkey

[6] https://www.crowdstrike.com/threat-intelligence/threat-ai/


Lyrie.ai Cyber Research Division

Lyrie Verdict

Lyrie's autonomous defense layer flags this class of exposure the moment it surfaces — no signature update required.