Digests

31 Results / Page 1 of 4

Background

todayJune 11, 2025

  • 339
close

MCP Security + Digests ADMIN

MCP Security Digest — June 2025

MCP Security is a top concern for anyone building Agentic AI systems. The Model Context Protocol (MCP) connects tools, agents, and actions. It plays a role similar to TCP/IP—but for autonomous workflows. If MCP is compromised, the entire agent stack is at risk. Attackers can inject prompts, hijack tools, and ...

todayJune 9, 2025

  • 131
close

Secure AI Weekly + Digests ADMIN

Towards Secure AI Week 22 — Testing the Limits of Guardrails and Autonomy

AI systems aren’t just generating answers—they’re taking action, reasoning independently, and connecting to real-world systems. This week’s stories highlight how current defenses fail to address these expanded capabilities, revealing critical blind spots in identity management, cross-agent communication, and cloud-based safety infrastructure. From one-shot jailbreaks and latent-level exploits to insecure identity ...

todayJune 2, 2025

  • 46
close

Secure AI Weekly + Digests ADMIN

Towards Secure AI Week 21 — From Reactive Defense to Capability-Aware AI Red Teaming

AI systems are no longer just responding to prompts — they’re acting, adapting, and making decisions. This week’s stories reveal how traditional security tools like SIEM, firewalls, and EDR fail to protect GenAI and Agentic AI systems, and why new approaches like continuous AI Red Teaming, identity enforcement, and jailbreak ...

todayMay 26, 2025

  • 146
close

Secure AI Weekly + Digests ADMIN

Towards Secure AI Week 20 — Identity, Jailbreaks, and the Future of Agentic AI Security

This week’s stories highlight the rapid emergence of new threats and defenses in the Agentic AI landscape. From OWASP’s DNS-inspired Agent Name Service (ANS) for verifying AI identities to real-world exploits like jailbreakable “dark LLMs” and prompt-injected assistants like GitLab Duo, the ecosystem is shifting toward identity-first architecture and layered ...

todayMay 19, 2025

  • 151
close

Secure AI Weekly + Digests ADMIN

Towards Secure AI Week 19 —  AI Agents Under Attack, Evaluation Becomes Strategy

This week’s stories highlight a critical evolution in AI risk: the shift from isolated agent failures to system-level compromise in Agentic AI architectures and memory-based applications. From Princeton’s demonstration of cryptocurrency theft via false memory injection to Fortnite’s AI Darth Vader being manipulated into swearing within an hour of launch, ...

todayMay 12, 2025

  • 175
close

Secure AI Weekly + Digests ADMIN

Towards Secure AI Week 18 — LLM Jailbreaks Hit New Highs, AI Security Market Accelerates

As LLMs become embedded across enterprise applications, new red-teaming research shows jailbreak success rates surpassing 87% on models like GPT-4—even under safety-aligned settings. Techniques such as multi-turn roleplay, token-level obfuscation, and cross-model attacks continue to outpace current safeguards. Meanwhile, insider misuse and unfiltered GenAI outputs pose growing risks, prompting calls ...

todayMay 7, 2025

  • 22
close

MCP Security + Digests ADMIN

MCP Security Digest — May 2025

MCP Security is a top concern for anyone building Agentic AI systems. The Model Context Protocol (MCP) connects tools, agents, and actions. It plays a role similar to TCP/IP—but for autonomous workflows. If MCP is compromised, the entire agent stack is at risk. Attackers can inject prompts, hijack tools, and ...

todayMay 5, 2025

  • 141
close

Secure AI Weekly + Digests admin

Towards Secure AI Week 17 — AI Guardrails Under Pressure as Jailbreaking Techniques Advance

Enterprise use of generative AI is expanding, but so is the sophistication of attacks targeting these systems. New jailbreak methods are achieving nearly 100% success rates, even on well-aligned models like GPT-4 and Llama3, while recent research exposes vulnerabilities in memory, prompt interpretation, and cross-tool coordination protocols like MCP. At ...

todayApril 28, 2025

  • 141
close

Secure AI Weekly + Digests admin

Towards Secure AI Week 16 — Can Your AI Agents Really Coordinate Safely?

As generative AI adoption accelerates, so do the security challenges that come with it. New research shows that even advanced large language models (LLMs) can be jailbroken with evolving techniques, while multi-agent AI systems introduce fresh risks at the communication and coordination layers. Cybercriminals are also scaling attacks using GenAI ...