
Towards Secure AI Week 9 – Exploiting AI Weaknesses
Researchers Jailbreak 17 Popular LLM Models to Reveal Sensitive Data GBHackers, March 7, 2025 Researchers from Palo Alto Networks’ Threat Research Center have discovered that 17 popular generative AI (GenAI) applications are vulnerable to jailbreaking techniques, allowing users to bypass safety protocols. By using both single-turn and multi-turn strategies, attackers ...