The GenAI security landscape entered 2026 with urgent warnings from global agencies and breakthrough research. CISA and international partners released comprehensive frameworks for securing AI in critical infrastructure, while researchers discovered that AI-generated code now shows 1.7x higher bug density than human-written code. Major stakeholders, including the UK NCSC and OpenAI, have recently issued sobering assessments suggesting that prompt injection might be an inherent, intractable property of LLM architectures. This month’s collection focuses on the industry’s pivot toward rigid verification and technical isolation as the only viable paths for securing enterprise generative AI deployments.
Statistics
Total resources: 107
Category breakdown:
GenAI security resources:
GenAI vulnerability
Oyster backdoor resurfaces: Analyzing the latest SEO poisoning attacks
The Oyster backdoor has resurfaced with SEO poisoning attacks targeting AI systems. Attackers are leveraging search optimization to compromise AI infrastructure.
ChatGPT Atlas vulnerable to context poisoning attacks
Reports indicate that ChatGPT Atlas systems are vulnerable to context poisoning. This attack vector manipulates the context window to alter model behavior.
AI security breach: Diffusion jailbreak threats rise
New diffusion jailbreak attacks are circumventing AI model safety guardrails in image generation models. This trend marks a growing sophistication in attacking multimodal AI.
The emerging threat of AI poisoning
This report analyzes the threat of data poisoning targeting AI training pipelines. It details how compromised data can corrupt model outputs.
It only takes a handful of samples to poison any size LLM
Anthropic researchers demonstrated that small numbers of samples can poison LLMs of any size. This reveals a critical fragility in data integrity controls.
AI models block 87% of single attacks, but just 8% when attackers persist
While AI models block most single attacks, they fail against persistent multi-step strategies. This reveals a significant effectiveness gap in adversarial defense.
LLMs and code security: Study reveals vulnerabilities in AI-generated code
A study reveals that LLM-generated code contains systematic vulnerabilities. These range from input validation bypasses to severe authentication flaws.
LangChain AI vulnerability exposes millions of apps
A critical vulnerability in LangChain exposes millions of apps to remote code execution. The flaw involves unsafe deserialization in the popular framework.
Researcher uncovers 30+ flaws in AI coding tools
New research has uncovered over 30 security flaws in popular AI coding tools. These vulnerabilities pose risks to the development environments they are integrated with.
Critical React2Shell flaw added to CISA KEV after weaponization
The React2Shell vulnerability has been added to CISA’s Known Exploited Vulnerabilities list. This flaw in AI-enabled systems is now being actively weaponized.
Critical flaws found in AI development tools IDE attack chain
An IDE attack chain has exposed critical flaws in AI development tools. These vulnerabilities can expose developers to data theft and remote code execution.
Gemini Enterprise no-click flaw exposes sensitive data
A no-click flaw in Gemini Enterprise could expose sensitive data without user interaction. This highlights the severity of passive vulnerabilities in enterprise AI tools.
Membership privacy risks of Sharpness Aware Minimization
Analysis of Sharpness Aware Minimization reveals privacy risks where membership info can be inferred. The study documents gradient-based leakage in training optimization methods.
Best of 2025: AI-generated code packages can lead to SlopSquatting threat
AI-generated packages are enabling SlopSquatting attacks, exploiting naming similarities. This demonstrates how AI-assisted supply chains can be tricked into downloading compromised code.
Worrying flaws already discovered in Googles Antigravity IDE
Google’s Antigravity IDE contains flaws that could expose sensitive environments. The discovery underscores the risks inherent in new AI-powered development tools.
Prompt injection attacks: The most common AI exploit in 2025
Direct prompt injection has become the most prevalent AI exploit in 2025, with a massive increase in success rates. The post provides a technical breakdown of techniques and defense strategies.
Adversarial image attacks (AI prompt injection vector)
Demonstrations show how adversarial image attacks use embedded malicious text to bypass defenses. This vector exploits multimodal processing to deliver payloads via images.
Indirect prompt injection attacks: Hidden AI risks
Crowdstrike details indirect prompt injection where external data poisons model behavior. It outlines detection and architectural defenses specifically for RAG systems.
OpenAI says AI browsers may always be vulnerable to prompt injection attacks
OpenAI research indicates that AI browsers may face permanent vulnerabilities due to architectural constraints. This suggests a need for defense-in-depth rather than relying on a single fix.
Prompt injection: The Achilles heel of AI assistants in the enterprise
This article examines prompt injection attacks as a critical security threat to enterprise assistants. It explains the attack methods and their impact on system integrity.
Researchers claim ChatGPT has security flaws including HackedGPT attack chain
The HackedGPT attack chain reveals multiple prompt injection vulnerabilities in ChatGPT-4o. It serves as a stark reminder of the complexity of securing conversational agents.
The hidden flaw in LLM safety: Translation as a jailbreak
Translation tasks act as a hidden jailbreak mechanism bypassing standard safety measures. The technique exploits the model’s multilingual capabilities to evade filters.
The risk of prompt injection: Your AI copilots can be hacked with words
This article explains how prompt injection exploits AI copilots to bypass controls. It details specific attack vectors used against assistants.
GenAI defense
Data poisoning threats: Protect AI with Cybarium
This resource outlines defense strategies against data poisoning in AI systems. It focuses on maintaining the integrity of training datasets.
A defense against data poisoning attacks in fine-tuned models
Researchers propose a defense mechanism specifically for fine-tuned LLMs. It includes detection techniques to identify malicious training data before model corruption.
Spectral sentinel: Scalable byzantine-robust learning
The Spectral Sentinel framework enables Byzantine-robust distributed learning at scale. It allows for secure collaborative training even with adversarial participants.
ArcGen: Generalizing neural backdoor detection
The ArcGen framework presents a method for generalizable neural backdoor detection. It uses spectral analysis to identify poisoned models across diverse architectures.
Jailbreak-proof AI security: Why zero trust beats guardrails
New analysis suggests that zero trust architecture outperforms traditional guardrails when defending against jailbreaks. The approach treats all inputs as potentially malicious, requiring verification at every boundary.
Harden your AI systems: Applying industry standards
RedHat provides guidance on hardening AI system deployments using established industry standards. Key topics include containerization, isolation, and compliance patterns.
How to build a secure generative AI architecture
This article offers guidance on designing secure GenAI architectures, covering data pipelines and output filtering. It integrates defensive patterns throughout the system lifecycle.
FedWiLoc: Federated learning for privacy-preserving WiFi localization
Researchers propose a federated learning approach for WiFi localization that removes the need for centralized data. It enables collaborative training while protecting sensitive location patterns.
S-AI-AntiHallucination: A bio-inspired and confidence-aware defense
This paper introduces a bio-inspired defense mechanism designed to prevent LLM hallucinations. The approach uses confidence-aware techniques to improve output reliability.
A secure threat detection framework for federated learning
This framework provides threat detection mechanisms specifically for federated learning environments. It aims to prevent model poisoning and inference attacks in distributed systems.
Real-time threat detection for post-quantum AI inference
A new framework offers real-time threat detection for post-quantum AI inference. It protects quantum-vulnerable models from emerging cryptographic attacks.
Israeli AI startup protects workplace chats from phishing
An Israeli startup has developed AI-based threat detection technology for workplace chats. The solution monitors conversational flows to identify phishing and social engineering.
Guarding machine learning hardware against physical attacks
This analysis covers physical side-channel attacks targeting ML hardware. It proposes hardware-level countermeasures for protecting model inference on edge devices.
Decentralized IoT security via deep federated learning
Researchers have developed an optimized federated learning framework for IoT security. It enables distributed threat detection while maintaining data privacy.
AI-based traffic modeling for network security and privacy
This paper proposes AI-based traffic modeling to enhance network security and privacy. The method utilizes machine learning for advanced pattern recognition in traffic flows.
Continuously hardening ChatGPT Atlas against prompt injection
OpenAI documents their iterative approach to hardening Atlas against prompt injection attacks. The post details practical improvements made through rigorous prompt engineering.
Adversarial image attacks (AI prompt injection vector)
This guide focuses on defending against adversarial image attacks in multimodal systems. It covers essential detection and validation techniques for non-text inputs.
ARGUS: Defending against multimodal indirect prompt injection attacks
Researchers propose the ARGUS defense framework to combat multimodal indirect prompt injections. The paper details how to secure systems against attacks hidden in non-text media.
Cloud security in 2026: Securing generative AI on AWS and Google
This guide covers securing generative AI deployments on major cloud platforms. It outlines specific defense strategies for AWS and Google Cloud environments.
OpenAI strengthens ChatGPT block prompt injection attacks
OpenAI has implemented stronger defenses in ChatGPT to detect and filter malicious input patterns. This update improves resilience against direct prompt attacks.
GenAI security research
Persistent backdoor attacks under continual fine-tuning
Research shows that backdoor attacks can survive multiple rounds of fine-tuning. This persistence poses a significant challenge for sanitizing corrupted models.
Threats and defenses for large language models: A survey
This academic survey synthesizes the current research on attack techniques and emerging threat vectors. It serves as a foundational overview of the AI security landscape.
The emerged security and privacy of LLM agent: A survey
A comprehensive survey covers the security and privacy issues specific to LLM agents. It addresses the unique risks of autonomous agentic systems.
Inside the AI supply chain: Security lessons from 10000 open source ML projects
This analysis examines supply chain security across thousands of open-source ML projects. It highlights systemic vulnerabilities in the development ecosystem.
Cryptographers show that AI protections will always have holes
Cryptographic analysis suggests that AI protection mechanisms have fundamental limitations. The research indicates that perfect security may be mathematically impossible.
Understanding generative AI in cybersecurity risks
This research examines the cybersecurity risks introduced by GenAI adoption in enterprises. It covers data exposure and lateral movement vectors enabled by integration.
Large language models as a bad security norm in the enterprise
A new paper argues that widespread LLM adoption creates organizational risk despite good intentions. It analyzes how model limitations can lead to operational security failures.
Security and privacy challenges of large language models
This survey categorizes security and privacy challenges across the LLM lifecycle. It covers vulnerabilities found in training, deployment, and active usage phases.
Understanding (new) security issues across AI4Code use cases
Researchers analyze emerging security issues in AI-for-code systems. The paper identifies governance gaps and risks like code injection and model poisoning.
Analysis of vulnerabilities attacks and countermeasures in AI
A comprehensive analysis covers vulnerabilities and countermeasures across AI systems. The paper details exploit techniques and defense mechanisms for ML infrastructure.
AI-generated code quality data shows 1.7x higher bug density in 2025
New data reveals that AI-generated code has significantly higher bug density than human code. The study quantifies the security quality gap in automated development.
Securing large language models (LLMs) from prompt injection
Research addresses methods for securing LLMs from prompt injection. The study evaluates various detection and prevention strategies against direct and indirect attacks.
Fighting AI with AI: The rise of multi-LLM orchestrated cyber attacks
Researchers explore the rise of multi-LLM orchestrated attacks. These coordinated attacks leverage multiple models to increase sophistication and evasion.
How safe are AI-generated patches? A large-scale study
A large-scale study finds that AI-generated security patches often introduce new vulnerabilities. The findings question the reliability of fully automated code remediation.
Trust in LLM-controlled robotics: A survey of security considerations
This survey examines security considerations for LLM-controlled robotics. It highlights physical safety risks resulting from compromised models.
Vulnerability of large language models to prompt injection
JAMA research demonstrates LLM vulnerability to prompt injection within clinical contexts. It provides critical evidence of risks in high-stakes medical AI applications.
Security reports
Anthropic AI misuse by Chinese hackers: How to defend
Case study on how hackers misused Anthropic models for reconnaissance. It demonstrates defensive measures against the dual-use of AI capabilities.
Lenovo AI threat report: 65% of IT leaders admit their defenses are outdated against AI attacks
A Lenovo survey reveals that 65% of leaders lack confidence in their AI defenses. It identifies a significant gap between adoption and security readiness.
DryRun Security finds over 80 percent of LLM application risks undetected
A study reveals that 80% of LLM risks go undetected by traditional code scanners. It highlights the failure of legacy tools to catch AI-specific flaws.
AI coding tools exploded in 2025. The first security exploits arrive.
Fortune documents the emerging exploits targeting AI coding tools. The report covers initial vulnerability discoveries in these widely adopted systems.
Science & tech spotlight: Malicious use of generative AI
The GAO released a report analyzing malicious use cases of generative AI across sectors. It provides a survey of threat vectors and recommendations for responsible policy.
Top AI security incidents of 2025 revealed
Adversa AI’s analysis reveals attack patterns against agentic AI systems. The report synthesizes threat data showing the rapid evolution of exploits.
Palo Alto Networks warns that AI is driving a surge in cloud security risk
Palo Alto Networks warns of an AI-driven surge in cloud risks including lateral movement. The analysis categorizes these vulnerabilities and offers defense recommendations.
Cybersecurity trends: IBMs predictions for 2026
IBM’s cybersecurity predictions for 2026 address AI-driven attack trends. The report synthesizes emerging threats and the requirements for organizational preparedness.
AI security trends 2026: Deepfakes, agents & LLM red teaming
Practical DevSecOps outlines key trends for 2026, including agentic AI risks. The outlook emphasizes the importance of intensive red-teaming initiatives.
Prompt injection attacks might never be properly mitigated – UK NCSC warns
The UK NCSC warns that prompt injection attacks might never be fully mitigated. The report suggests organizations must plan for these risks as a permanent feature.
ThreatsDay bulletin: Stealth loaders AI chatbot flaws
The ThreatsDay Bulletin covers stealth loaders and chatbot flaws that enable threat delivery. It highlights multiple security issues in current AI implementations.
OWASP says prompt injection is the #1 LLM threat for 2025
OWASP has identified prompt injection as the top threat for LLMs in 2025. This ranking reflects the high impact and frequency of prompt-based attacks.
Article
Spy vs spy: How GenAI is powering defenders and attackers
Talos Intelligence compares how GenAI powers both defenders and attackers. The analysis highlights that the advantage depends on organizational detection capabilities.
Adversarial poetry and the efficacy of AI guardrails
This article analyzes adversarial poetry attacks and the effectiveness of current AI guardrails. It explores how creative language can bypass filters.
Adversarial attacks on large language models
This article examines adversarial attacks targeting LLMs and their inherent vulnerabilities. It explores the mechanics of how these models can be manipulated.
Traditional security frameworks leave organizations exposed to AI
Analysis indicates that traditional frameworks are inadequate for AI-specific risks. The article advocates for adopting AI-native security approaches.
AI-powered cybersecurity: Applications and risks
This post discusses the applications and risks of using AI-powered tools in cybersecurity. It balances the benefits of automation against potential new weaknesses.
How AI transforms software supply chain security
This article examines how AI is transforming supply chain security through automated detection. It discusses both the defensive improvements and the new attack vectors created.
AI inSecurity: We’re making the same mistakes again
This article argues that the industry is repeating historical security mistakes with AI. It advocates for applying traditional threat modeling disciplines to new systems.
Think AI can replace your security team? Think again.
Analysis suggests that AI cannot fully replace human security teams due to context limitations. It argues for AI as a complement to, rather than a replacement for, professionals.
Pen testers accused of blackmail after reporting Eurostar chatbot flaws
Researchers reporting chatbot vulnerabilities at Eurostar faced blackmail accusations. The incident illustrates the growing tension between security researchers and AI service providers.
OpenAIs ChatGPT Atlas browser faces security scrutiny
OpenAI’s ChatGPT Atlas browser extension is under scrutiny for potential privacy and security risks. The concerns focus on vulnerabilities introduced by deep web integration.
Defense frameworks
CISA, Australia, and partners author joint guidance on securely integrating AI
CISA and international partners have released a comprehensive framework for secure AI integration within critical infrastructure. It addresses general risks and best practices applicable across various threat models.
NIST AI cybersecurity framework: Defense against real threats
The updated NIST framework adapts foundational cybersecurity principles specifically for AI systems. It covers governance, risk management, and technical controls required for robust defense.
MITRE ATLAS framework 2025 – Guide to securing AI systems
The MITRE ATLAS 2025 guide provides an updated taxonomy of AI security threats and defense strategies. It establishes a necessary common language for discussing AI system security.
A CISSP-inspired AI security approach
ISC2 has established a new AI security competency framework defining core professional knowledge areas. It offers a structured approach for organizations building internal expertise.
Global cyber agencies issue AI security guidance for critical infrastructure OT
International cyber agencies have issued specific guidance for AI in OT environments. The document establishes baseline security requirements for critical infrastructure operations.
Unpacking the OWASP AI testing guide 2025
The OWASP AI Testing Guide 2025 outlines a methodology for identifying vulnerabilities. It offers structured validation steps aligned with the modern development lifecycle.
CISA publishes security guidance for using AI in OT
CISA has published official security guidance for integrating AI into operational technology. It covers best practices and defense frameworks for industrial environments.
Solving the unique challenges of AI security and delivery
F5 presents a comprehensive framework for addressing authentication, encryption, and threat detection. It addresses the unique requirements of delivering AI applications securely.
Threat modelling
From phishing to bias: Study maps the hidden threats
An academic study maps the comprehensive threat landscape for LLMs. It reveals the interconnected nature of bias, data poisoning, and phishing.
Adversarial ML taxonomy
A comprehensive taxonomy of adversarial threats categorizing attacks by target and method. This framework helps in understanding and classifying adversarial machine learning.
LLM vulnerabilities: Why AI models are the next big attack target
This analysis explains why LLMs are high-value targets compared to traditional software. It covers model extraction and membership inference vulnerabilities.
Supply chain data attacks: Trust no data
This post emphasizes supply chain data risks where artifacts are compromised during development. It addresses the trust challenges in AI-mediated pipelines.
GenAI security 101
AI security risks every business must know about
This article synthesizes critical AI security risks for business awareness. It addresses model poisoning, supply chain vulnerabilities, and privacy threats.
What is a prompt injection attack and how it works
Norton’s guide explains prompt injection mechanics with practical defense techniques. It covers attack patterns and mitigation for organizations.
What is a prompt injection attack?
An educational overview of prompt injection tactics and mechanisms. This is a solid primer for understanding the basics of this vector.
Attack
Aurascape researchers expose new AI attack that sends travelers to scam airline support
Researchers expose an AI-based social engineering attack targeting travelers. The exploit redirects users to scam support centers using generated content.
Exploiting false positives for denial-of-service attacks
Attackers can mount denial-of-service attacks by exploiting false positives in AI detection. This technique demonstrates computational vulnerabilities in safety mechanisms.
Datasets
Language model security database
The Language Model Security Database documents known vulnerabilities and attack vectors. It is a key resource for researchers tracking LLM flaws.
TeleAI-Safety: A comprehensive LLM jailbreaking benchmark
TeleAI-Safety is a comprehensive benchmark for evaluating jailbreak attacks and defenses. It provides a framework for testing model robustness.
Resource collections on generative AI
10 best AI security tools for LLM protection 2025
A curated list of defense tools and frameworks for LLM protection. This resource covers the top options for securing GenAI deployments.
PromptIntel – AI security threat intelligence
PromptIntel provides threat intelligence focused on monitoring prompt-based attacks. The platform helps in threat hunting specifically for LLM applications.
Attack techniques
Poetry jailbreaks hit 100 percent on some models
A new poetry-based adversarial attack achieves 100% success on some LLMs. The technique uses creative writing to bypass standard filters.
Researchers use poetry to jailbreak AI models
Researchers discovered that jailbreaking via poetry bypasses safety mechanisms by exploiting artistic contexts. This demonstrates a novel avenue for exploitation.
Test LLM resilience with new Cymulate attack scenarios
Cymulate’s platform allows you to test LLM resilience through realistic prompt injection scenarios. The tool enables continuous security testing for AI applications.
AI-based web app scanner ‘Rogue’ uses OpenAI for analysis
Rogue is a scanner that uses OpenAI models to analyze web vulnerabilities. It represents the augmentation of traditional scanning with AI reasoning.
GenAI red teaming
Learning-based automated adversarial red-teaming for language models
This paper presents automated red-teaming using AI to systematically generate attacks. The technique helps in identifying vulnerabilities at scale.
Security incident
12000 leaked secrets in AI training data
12,000 secrets were leaked through AI training data, creating widespread risks. This incident demonstrates the severe impact of data poisoning through exposure.
Training materials
AI security audits for LLMs lessons from Anthropic
Anthropic shares lessons on security audits of LLMs, emphasizing red-teaming and safety testing. The article documents their approach to comprehensive assessment.
Videos
Adversarial AI: The new symmetric threat landscape
This video examines adversarial AI as an emerging symmetric threat. It explores the changing dynamics of the cybersecurity landscape.
Moving beyond the linguistic firewall
The The sheer volume of vulnerabilities discovered this month makes one thing clear: passive observation is no longer sufficient; linguistic safety filters or “soft” guardrails are not enough. With persistent attackers achieving success over 90% of the time, organizations must move toward a zero-trust approach in production AI systems, implementing sandboxing for any AI agent with tool access and treating every model output as an unauthenticated command until verified by a secondary validation layer. The era of tightly controlled model agency has begun.
Subscribe for updates
Stay up to date with what is happening! Plus, get a first look at news, noteworthy research, and the worst attacks on AI — delivered right to your inbox.