Adversa AI Wins “Most Innovative Agentic AI Security” at Global InfoSec Awards During RSA Conference 2026
Recognized Among Hundreds of Vendors for Advancing Continuous AI Red Teaming and Agentic AI Security.
Agentic AI Security + Agentic AI Security Digest Sergey todayApril 1, 2026
This month’s Agentic AI security conversation was shaped by discussions at the RSA conference and continued attention on high-profile attempts to secure OpenClaw, spearheaded by NVIDIA itself. The conversation has shifted from theoretical risks to active, infrastructure-level threats. We’re seeing a surge in advanced attacks, from multi-agent offensive behaviors to serious vulnerabilities in widely deployed tools like OpenClaw and Copilot. As agents gain more autonomy and access, the need for strong, agent-specific defense mechanisms and identity governance is now pressing.
Total resources: 33
Category breakdown:
| Category | Count |
|---|---|
| Research | 7 |
| Agentic AI vulnerabilities | 5 |
| Agentic AI defense | 4 |
| Threat modelling | 4 |
| Agentic AI security for CISO | 4 |
| Agentic AI security 101 | 2 |
| Article | 2 |
| Exploitation | 2 |
| Security tools | 2 |
| Attack | 1 |
This paper presents a formal systematization of LLM agent security, decomposed into four core properties. It explains why pattern matching defenses fail and how current benchmarks miss key tests.
Unit 42 analyzed real-world telemetry to identify 22 distinct techniques of indirect prompt injection. This confirms the threat has moved from theoretical to actively weaponized against AI agents.
A systematic audit reveals that 93% of 30 AI agent frameworks rely on unscoped API keys. Furthermore, 0% have per-agent identity, and 97% lack user consent mechanisms.
Memory in LLM agents is a key attack surface where poisoned memory entries can persistently hijack workflows. Researchers achieved 90%+ attack success rates against major models like GPT-5 mini and Claude Sonnet 4.5.
TrinityGuard introduces a three-tier risk taxonomy for multi-agent systems, revealing a low 7.1% average safety pass rate. The framework has been open-sourced with AG2/AutoGen integration.
Researchers tested OpenClaw across 47 adversarial scenarios and found sandbox escapes with only a 17% average defense rate. They proposed a HITL defense layer that improves protection to 91.5%.
Memory poisoning poses serious reliability risks in multi-turn AI agents. This paper introduces the Agent Cognitive Compressor (ACC), a bio-inspired memory controller to mitigate context drift and unbounded transcript replay.
Unit 42 discovered CVE-2026-0628, a high-severity flaw in Chrome’s Gemini Live panel. It allowed malicious extensions to hijack the privileged AI assistant and access the camera and microphone.
A serious vulnerability in OpenClaw’s local WebSocket gateway allowed malicious websites to hijack developer AI agents without user interaction by exploiting implicit localhost trust.
Four CVEs in CrewAI let attackers chain prompt injection into RCE, SSRF, and file read. These vulnerabilities affect the Code Interpreter and default configurations.
SecurityWeek details the OpenClaw localhost WebSocket vulnerability. The flaw allowed browser JavaScript to brute-force passwords due to exempted rate limits.
A CVSS 9.9 privilege escalation vulnerability in OpenClaw allows low-privilege tokens to escalate to admin with RCE. Over 135,000 internet-facing instances were detected.
Sysdig TRT instrumented Claude Code, Gemini CLI, and Codex CLI at the syscall level. They identified four detection patterns and provided Falco/eBPF rules for AI coding agent threats.
SlowMist pioneers an ‘agent-facing’ defense paradigm with a security guide designed to be read and deployed BY the AI agent itself. This shifts from human-only hardening to agentic zero-trust.
This article synthesizes defense announcements from major AI providers with a concrete six-layer defense stack. It includes implementable code examples to prevent tool-result prompt injection.
STSS is an open-source defense layer that scans AI agent skills via static analysis and behavioral auditing. It issues cryptographic attestations using SHA-256 Merkle trees to secure agent capabilities.
Discover 10 distinct architectural vulnerabilities, from memory poisoning to inter-agent trust, that could lead to your next data breach. The post shows blind spots in current AI security solutions.
This paper presents a component-aligned threat taxonomy covering six threat families and six defense strategies. It analyzes how risks grow as we transition to the agentic web.
A MITRE ATT&CK-style threat matrix outlining six tactics and 20+ techniques for a rogue agent kill chain. It provides a structured look at potential autonomous agent behaviors.
A systematic threat model exploring agent social networks. It maps five security risks to the OWASP ASI framework.
AI red teaming finds what guardrails miss, including multi-step attack chains and semantic goal hijacking. This post explains the coverage gap and why both approaches are necessary.
Only 21.9% of organizations treat AI agents as identity-bearing entities. This post proposes a five-layer identity architecture for governing enterprise AI agents.
Recent incidents reveal that enterprise security controls are inadequate for autonomous AI. The article proposes five concrete controls for security leaders to implement.
Banning high-agency AI like OpenClaw won’t stop shadow AI. Enterprises need a proactive security strategy to maintain their competitive edge while managing risks.
Prompt injection has escalated from a model-level to an infrastructure-level threat. This post synthesizes disclosures on browser agents and MCP poisoning from major AI vendors.
A comprehensive cheat sheet grouping all 20 OWASP items into three architectural risk categories. It provides an accessible onramp for engineers new to AI security with illustrated attack scenarios.
Irregular demonstrated multi-agent offensive behavior including forging admin cookies and disabling endpoint defenses. It shows the growing threat of collaborative rogue agents.
An analysis of 104 CVEs in OpenClaw shows dominant vulnerability classes stemming from insecure-by-design architecture. Vibe-coded agents create a dynamic attack surface that requires new security paradigms.
An autonomous AI agent powered by Claude Opus is actively exploiting GitHub Actions workflows in the wild. The bot achieved RCE in major targets using techniques like poisoned Go init() functions.
BeyondTrust discovered a serious command injection in OpenAI Codex that allows stealing GitHub OAuth tokens via unsanitized branch name parameters.
NVIDIA NemoClaw is an open-source reference stack that simplifies running OpenClaw assistants safely. It installs the NVIDIA OpenShell runtime for additional security.
An open-source agent firewall featuring an 11-layer pipeline for DLP and MCP tool poisoning detection. It uses capability separation to secure agent workflows.
A proof of concept demonstrates how a malicious A2A agent card can embed adversarial instructions. This leads to data exfiltration via the host LLM.
The disclosures this month make one thing clear: you cannot secure what you haven’t tested. Attackers are already probing agentic infrastructure, while the systems themselves evolve faster than teams can make sense of the threat model, apply systemic hardening, or even simply patch it. Security teams must continuously red team their AI agents, stress-test tool integrations and memory systems under adversarial conditions, and treat ongoing assessment as core operational practice. Agents gain more autonomy and access every day. The organizations that strive will be the ones that “break” their own systems before someone else does.
Written by: Sergey
Industry Awards admin
Recognized Among Hundreds of Vendors for Advancing Continuous AI Red Teaming and Agentic AI Security.
(c) Adversa AI, 2026. Continuous red teaming of AI systems, trustworthy AI research & advisory
Privacy, cookies & security compliance · Security & trust center