Top GenAI Security Resources — December 2025

GenAI Security + GenAI Security Digest Sergey todayDecember 5, 2025

Background
share close

GenAI Security: Essential Resources for December 2025

Generative AI has rapidly become ubiquitous in business applications, and the installed base of AI assistants already exceeds one billion users. Security considerations for this wide adoption range from sophisticated prompt-injection attacks to novel side-channel vulnerabilities, and the threat landscape for AI systems continues to evolve at an unprecedented pace. This digest compiles the most important GenAI security resources from December 2025 to help security professionals stay ahead of emerging risks.

Statistics

This digest covers 26 top GenAI security resources across 11 categories:

  • Exploitation Techniques: 8 resources (31%)
  • Security Tools: 5 resources (19%)
  • Security Research: 2 resources (8%)
  • Attack Techniques: 2 resources (8%)
  • CISO Resources: 1 resource (4%)
  • Security Incident: 1 resource (4%)
  • GenAI Vulnerability: 1 resource (4%)
  • Defense Mechanisms: 1 resource (4%)
  • Red-Teaming: 1 resource (4%)
  • Security Training: 1 resource (4%)
  • GenAI Security 101: 1 resource (4%)
  • Safety Dataset: 1 resource (4%)
  • Security Report: 1 resource (4%)

Content

Exploitation Techniques

Zero-G, Zero Trust: How Antigravity Floats Away with Your Secrets

This article explores a novel technique called Antigravity that exploits trust boundaries in AI systems. It demonstrates how attackers can bypass security controls to exfiltrate sensitive information from AI-powered applications.

HackedGPT: Novel AI Vulnerabilities Open the Door for Private Data Leakage

Tenable researchers uncover new AI vulnerabilities that enable private data leakage. The findings highlight critical security gaps in popular AI implementations that organizations must address.

PromptJacking: The Critical RCEs in Claude Desktop That Turn Questions Into Exploits

Security researchers demonstrate critical remote code execution vulnerabilities in Claude Desktop through prompt manipulation. The attack chain shows how innocent-looking prompts can be weaponized.

AI/LLM Hacking – Part 5 – Improper Output Handling

This tutorial covers improper output handling vulnerabilities in LLMs. It provides practical examples of how attackers exploit these weaknesses and offers mitigation strategies.

Prompt Injection Variant Lets Hackers Exfiltrate Data from Claude APIs

A new prompt injection variant enables data exfiltration from Claude APIs. The technique bypasses existing security measures and poses risks to applications using Claude’s API endpoints.

Persistent XSS via Prompt Injection with Client-Side Exfiltration in DeepSeek-V3

A GitHub issue documents a persistent XSS vulnerability in DeepSeek-V3 triggered through prompt injection. The attack enables client-side data exfiltration from affected systems.

Introducing InjectPrompt Companion

A new tool designed to help security professionals test and analyze prompt injection vulnerabilities. It provides automated testing capabilities for identifying AI security weaknesses.

Critical AI Safety Jailbreak Alert

A LinkedIn post reveals a new jailbreak technique targeting Claude Haiku. The discovery highlights ongoing challenges in maintaining AI safety guardrails against adversarial attacks.

Security Tools

llm-security-framework

An open-source security framework for testing and securing LLM deployments. It provides a comprehensive toolkit for identifying vulnerabilities in AI systems.

Security Testing for Production LLM Applications

A system prompt benchmark tool for security testing production LLM applications. It helps organizations evaluate the robustness of their AI systems against common attack vectors.

Whisper Leak Tool

An open-source implementation of the Whisper Leak side-channel attack. Security researchers can use this tool to test their systems against this novel attack technique.

MLEBENCH-SUBVERSION

A benchmarking tool for evaluating AI system subversion attempts. It helps measure and improve the resilience of machine learning systems against adversarial manipulation.

Security Research

Chain-of-Thought Hijacking

Academic research exploring how attackers can manipulate chain-of-thought reasoning in LLMs. The paper demonstrates techniques to hijack the reasoning process for malicious purposes.

Adversarial Poetry as a Universal Single-Turn Jailbreak Mechanism in Large Language Models

Researchers discover that adversarial poetry can serve as a universal jailbreak technique. This single-turn attack method bypasses safety measures across multiple LLM platforms.

Attack Techniques

Mind-reading Claude AI’s Complete System Prompt

A detailed exploration of techniques used to extract Claude AI’s system prompts. The article reveals methods attackers use to uncover hidden instructions in AI systems.

Multi-Layered Persona Injection with Rule Override Attempt

Documentation of a sophisticated prompt injection technique using multi-layered personas. The attack attempts to override AI system rules through carefully crafted identity manipulation.

CISO Resources

The Big Idea: Security Assurance is NOT Just QA for AI!

Google Cloud’s CISO blog explains why AI security assurance requires a fundamentally different approach than traditional QA. Essential reading for security leaders implementing AI governance.

Security Incident

ShadowRay 2.0: Attackers Turn AI Against Itself in Global Campaign that Hijacks AI Into Self-Propagating Botnet

Oligo Security uncovers a global attack campaign where AI systems are hijacked to form self-propagating botnets. The incident demonstrates how attackers weaponize AI infrastructure at scale.

GenAI Vulnerability

Whisper Leak: A Novel Side-Channel Attack on Remote Language Models

Microsoft Security reveals a novel side-channel attack targeting remote language models. The vulnerability allows attackers to extract sensitive information through timing analysis.

Defense Mechanisms

The Evolution of AI Security: Why Secure AI by Design Matters

Palo Alto Networks outlines the importance of building security into AI systems from the ground up. The article provides a framework for implementing secure AI design principles.

Red-Teaming

Build a Pro AI Hacking Lab: Microsoft Red Teaming Playground Setup

A video tutorial on setting up Microsoft’s AI red teaming playground. It provides step-by-step guidance for security professionals building AI testing environments.

Security Training

Understanding Prompt Injection Techniques, Challenges, and Advanced Escalation

A comprehensive training course by Brian Vermeer covering prompt injection fundamentals. It explores techniques, real-world challenges, and advanced escalation methods.

GenAI Security 101

The Beginner’s Guide to AI Security (and Why It Matters) — Part 1

An accessible introduction to AI security concepts for newcomers. It explains fundamental risks and why organizations must prioritize AI security in their strategies.

Safety Dataset

Real-World LLM Misuse Datasets and Taxonomies

A curated collection of real-world LLM misuse datasets and classification taxonomies. Valuable resource for researchers studying AI safety and developing protective measures.

Security Report

The AI Security Blind Spot

Harness releases a comprehensive report on AI-native application security. It identifies critical blind spots organizations face when securing AI-powered applications.

Conclusion

The GenAI security landscape continues to evolve rapidly, with new attack vectors and defense mechanisms emerging constantly. The gap between attack innovation and organizational preparedness continues to widen, as evidenced by incidents like ShadowRay 2.0. Organizations can no longer treat AI security as an afterthought or extension of traditional cybersecurity — the time for proactive AI security strategies, red team exercises, and specialized defensive tools is now, before your AI systems become the next attack vector.

Stay vigilant, stay informed, and remember: in the rapidly evolving world of GenAI security, today’s innovative defense becomes tomorrow’s baseline requirement.

Subscribe for updates

Stay up to date with what is happening! Plus, get a first look at news, noteworthy research, and the worst attacks on AI — delivered right to your inbox.

    Written by: Sergey

    Rate it
    Previous post