Trusted AI Blog

342 Results / Page 1 of 38

Cyberpunk-style illustration of AI systems like ChatGPT, Claude, Grok, and Deepseek facing prompt injection attacks

todayMay 22, 2025

  • 87
close

Articles ADMIN

Prompt Injection Risks Interview: Are AIs Ready to Defend Themselves? Conversation with ChatGPT, Claude, Grok & Deepseek

Prompt injection remains one of the most dangerous and poorly understood threats in AI security. To assess how today’s large language models (LLMs) handle Prompt Injection risks, we interviewed ChatGPT, Claude, Grok, and Deepseek. We asked each of them 11 expert-level questions covering real-world attacks, defense strategies, and future readiness. ...

Abstract AI security background with glitch effects and shield symbol, representing trust and resilience in generative AI.

todayMay 19, 2025

  • 116
close

Secure AI Weekly ADMIN

Towards Secure AI Week 19 —  AI Agents Under Attack, Evaluation Becomes Strategy

This week’s stories highlight a critical evolution in AI risk: the shift from isolated agent failures to system-level compromise in Agentic AI architectures and memory-based applications. From Princeton’s demonstration of cryptocurrency theft via false memory injection to Fortnite’s AI Darth Vader being manipulated into swearing within an hour of launch, ...

Abstract AI security background with glitch effects and shield symbol, representing trust and resilience in generative AI.

todayMay 14, 2025

  • 86
close

Articles ADMIN

ETSI TS 104 223: 10 Security Insights Every CISO Needs

As AI systems rapidly integrate into critical infrastructure and enterprise workflows, their attack surfaces are expanding just as quickly. Consequently, traditional cybersecurity controls are no longer sufficient. To address this growing risk, the new ETSI TS 104 223 V1.1.1 (2025-04) — Securing Artificial Intelligence (SAI); Baseline Cyber Security Requirements for ...

Abstract AI security background with glitch effects and shield symbol, representing trust and resilience in generative AI.

todayMay 12, 2025

  • 146
close

Secure AI Weekly ADMIN

Towards Secure AI Week 18 — LLM Jailbreaks Hit New Highs, AI Security Market Accelerates

As LLMs become embedded across enterprise applications, new red-teaming research shows jailbreak success rates surpassing 87% on models like GPT-4—even under safety-aligned settings. Techniques such as multi-turn roleplay, token-level obfuscation, and cross-model attacks continue to outpace current safeguards. Meanwhile, insider misuse and unfiltered GenAI outputs pose growing risks, prompting calls ...

Abstract AI security background with glitch effects and shield symbol, representing trust and resilience in generative AI.

todayMay 5, 2025

  • 78
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 17 — AI Guardrails Under Pressure as Jailbreaking Techniques Advance

Enterprise use of generative AI is expanding, but so is the sophistication of attacks targeting these systems. New jailbreak methods are achieving nearly 100% success rates, even on well-aligned models like GPT-4 and Llama3, while recent research exposes vulnerabilities in memory, prompt interpretation, and cross-tool coordination protocols like MCP. At ...

todayApril 30, 2025

  • 505
close

Articles ADMIN

Agentic AI Security: Key Threats, Attacks, and Defenses

Artificial Intelligence has entered a new phase. No longer limited to generating text or analyzing data, AI systems can now take initiative. Meet Agentic AI—autonomous systems capable of making decisions, interacting with APIs, browsing the web, updating spreadsheets, sending emails, and executing code which means we need an Agentic AI ...

todayApril 28, 2025

  • 123
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 16 — Can Your AI Agents Really Coordinate Safely?

As generative AI adoption accelerates, so do the security challenges that come with it. New research shows that even advanced large language models (LLMs) can be jailbroken with evolving techniques, while multi-agent AI systems introduce fresh risks at the communication and coordination layers. Cybercriminals are also scaling attacks using GenAI ...