Trusted AI Blog

475 Results / Page 10 of 53

Background

todayMay 20, 2025

  • 336
close

Agentic AI Security + Review ADMIN

Microsoft’s Taxonomy of Failure Modes in Agentic AI Systems — TOP 10 Insights 

Based on Microsoft AI Red Team’s white paper “Taxonomy of Failure Modes in Agentic AI Systems”. Why CISOs, Architects & Staff Engineers Must Read Microsoft’s Agentic AI Failure Mode Taxonomy Agentic AI is moving from proof-of-concept to production faster than most security teams can update their threat models. In response, ...

todayMay 19, 2025

  • 169
close

Secure AI Weekly ADMIN

Towards Secure AI Week 19 —  AI Agents Under Attack, Evaluation Becomes Strategy

This week’s stories highlight a critical evolution in AI risk: the shift from isolated agent failures to system-level compromise in Agentic AI architectures and memory-based applications. From Princeton’s demonstration of cryptocurrency theft via false memory injection to Fortnite’s AI Darth Vader being manipulated into swearing within an hour of launch, ...

todayMay 14, 2025

  • 156
close

Review + GenAI Security ADMIN

ETSI TS 104 223: 10 Security Insights Every CISO Needs

As AI systems rapidly integrate into critical infrastructure and enterprise workflows, their attack surfaces are expanding just as quickly. Consequently, traditional cybersecurity controls are no longer sufficient. To address this growing risk, the new ETSI TS 104 223 V1.1.1 (2025-04) — Securing Artificial Intelligence (SAI); Baseline Cyber Security Requirements for ...

todayMay 12, 2025

  • 204
close

Secure AI Weekly ADMIN

Towards Secure AI Week 18 — LLM Jailbreaks Hit New Highs, AI Security Market Accelerates

As LLMs become embedded across enterprise applications, new red-teaming research shows jailbreak success rates surpassing 87% on models like GPT-4—even under safety-aligned settings. Techniques such as multi-turn roleplay, token-level obfuscation, and cross-model attacks continue to outpace current safeguards. Meanwhile, insider misuse and unfiltered GenAI outputs pose growing risks, prompting calls ...

todayMay 7, 2025

  • 73
close

MCP Security + MCP Security Digest ADMIN

MCP Security Digest — May 2025

MCP Security is a top concern for anyone building Agentic AI systems. The Model Context Protocol (MCP) connects tools, agents, and actions. It plays a role similar to TCP/IP—but for autonomous workflows. If MCP is compromised, the entire agent stack is at risk. Attackers can inject prompts, hijack tools, and ...

todayMay 5, 2025

  • 168
close

Secure AI Weekly admin

Towards Secure AI Week 17 — AI Guardrails Under Pressure as Jailbreaking Techniques Advance

Enterprise use of generative AI is expanding, but so is the sophistication of attacks targeting these systems. New jailbreak methods are achieving nearly 100% success rates, even on well-aligned models like GPT-4 and Llama3, while recent research exposes vulnerabilities in memory, prompt interpretation, and cross-tool coordination protocols like MCP. At ...

todayApril 28, 2025

  • 154
close

Secure AI Weekly admin

Towards Secure AI Week 16 — Can Your AI Agents Really Coordinate Safely?

As generative AI adoption accelerates, so do the security challenges that come with it. New research shows that even advanced large language models (LLMs) can be jailbroken with evolving techniques, while multi-agent AI systems introduce fresh risks at the communication and coordination layers. Cybercriminals are also scaling attacks using GenAI ...