Publications

55 Results / Page 5 of 7

todayJanuary 31, 2025

  • 18840
close

Research + LLM Security admin

DeepSeek Jailbreak’s

Deepseek Jailbreak’s In this article, we will demonstrate how DeepSeek respond to different jailbreak techniques. Our initial study on AI Red Teaming different LLM Models using various aproaches focused on LLM models released before the so-called “Reasoning Revolution”, offering a baseline for security assessments before the emergence of advanced reasoning-based ...

todayApril 2, 2024

  • 3802
close

Research + LLM Security admin

LLM Red Teaming: Adversarial, Programming, and Linguistic approaches VS ChatGPT, Claude, Mistral, Grok, LLAMA, and Gemini

Warning, Some of the examples may be harmful!: The authors of this article show LLM Red Teaming and hacking techniques but have no intention to endorse or support any recommendations made by AI Chatbots discussed in this post. The sole purpose of this article is to provide educational information and ...

todayNovember 15, 2023

  • 5342
  • 2
close

Research + LLM Security admin

What is Prompt Leaking, API Leaking, Documents Leaking in LLM Red Teaming

What is AI Prompt Leaking? Adversa AI Research team revealed a number of new LLM Vulnerabilities, including those resulted in Prompt Leaking that affect almost any Custom GPT’s right now.  Subscribe for the latest LLM Security news: Prompt Leaking, Jailbreaks, Attacks, CISO guides, VC Reviews, and more Step one. Approximate Prompt ...

todayNovember 1, 2023

  • 122
close

Article + LLM Security admin

White House Executive Order On Safe And Secure AI: A Need For External AI Red Teaming

Why is it important? In recognition of AI’s transformative potential and the associated challenges, President Biden has taken the decisive step of issuing an Executive Order geared toward ensuring AI evolves safely, securely, and in the best interest of all Americans. Given the expansive impacts of AI, it’s pivotal that ...

todayMarch 15, 2023

  • 31153
close

Research + LLM Security admin

GPT-4 Jailbreak and Hacking via RabbitHole attack, Prompt injection, Content moderation bypass and Weaponizing AI

GPT-4 Jailbreak is what all the users have been waiting for since the GPT-4 release. We gave it within 1 hour. Subscribe for the latest AI Jailbreaks, Attacks, and Vulnerabilities Today marks the highly anticipated release of OpenAI’s GPT-4, the latest iteration of the groundbreaking natural language processing and  CV ...

ChatGPT hacking

todayDecember 6, 2022

  • 6956
close

Research + LLM Security admin

ChatGPT Security: eliminating humanity and hacking Dalle-2 using a trick from Jay and Silent Bob

ChatGPT Security note: The authors of this article show ChatGPT hacking techniques but have no intention to endorse or support any recommendations made by ChatGPT discussed in this post. The sole purpose of this article is to provide educational information and examples for research purposes to improve the security and ...