Research

16 Results / Page 2 of 2

todayApril 2, 2024

  • 3802
close

Research + LLM Security admin

LLM Red Teaming: Adversarial, Programming, and Linguistic approaches VS ChatGPT, Claude, Mistral, Grok, LLAMA, and Gemini

Warning, Some of the examples may be harmful!: The authors of this article show LLM Red Teaming and hacking techniques but have no intention to endorse or support any recommendations made by AI Chatbots discussed in this post. The sole purpose of this article is to provide educational information and ...

todayNovember 15, 2023

  • 5342
  • 2
close

Research + LLM Security admin

What is Prompt Leaking, API Leaking, Documents Leaking in LLM Red Teaming

What is AI Prompt Leaking? Adversa AI Research team revealed a number of new LLM Vulnerabilities, including those resulted in Prompt Leaking that affect almost any Custom GPT’s right now.  Subscribe for the latest LLM Security news: Prompt Leaking, Jailbreaks, Attacks, CISO guides, VC Reviews, and more Step one. Approximate Prompt ...

todayMarch 15, 2023

  • 31153
close

Research + LLM Security admin

GPT-4 Jailbreak and Hacking via RabbitHole attack, Prompt injection, Content moderation bypass and Weaponizing AI

GPT-4 Jailbreak is what all the users have been waiting for since the GPT-4 release. We gave it within 1 hour. Subscribe for the latest AI Jailbreaks, Attacks, and Vulnerabilities Today marks the highly anticipated release of OpenAI’s GPT-4, the latest iteration of the groundbreaking natural language processing and  CV ...

ChatGPT hacking

todayDecember 6, 2022

  • 6956
close

Research + LLM Security admin

ChatGPT Security: eliminating humanity and hacking Dalle-2 using a trick from Jay and Silent Bob

ChatGPT Security note: The authors of this article show ChatGPT hacking techniques but have no intention to endorse or support any recommendations made by ChatGPT discussed in this post. The sole purpose of this article is to provide educational information and examples for research purposes to improve the security and ...

todayJune 30, 2019

  • 60
close

Research + Adversarial ML admin

Tricks of the trade: fooling identification models with perturbed audio, image and biometric input

Adversa is once again sharing the research that captured our interest. In June 2019 we marveled at the incredibly effective new adversarial model with a whopping 97% success rate, learned about Youtube’s copyright system and the potential of human eyes in authentication, and saw the effects of human biases on ...