Trusted AI Blog

475 Results / Page 12 of 53

Background

todayMarch 4, 2025

  • 113
close

Secure AI Weekly admin

Towards Secure AI Week 8 – Massive AI security breach

MITRE Releases OCCULT Framework to Address AI Security Challenges GBHackers, February 26, 2025 MITRE has launched the Offensive Cyber Capability Unified LLM Testing (OCCULT) framework, a structured approach designed to assess the potential security risks of large language models (LLMs) in cyberattacks. This initiative focuses on evaluating AI systems’ ability ...

todayFebruary 25, 2025

  • 147
close

Secure AI Weekly admin

Towards Secure AI Week 7 – OWASP for Agentic AI and more

Agentic AI – Threats and Mitigations OWASP, February 17, 2025 Agentic AI, driven by large language models (LLMs) and generative AI, is advancing rapidly, offering new capabilities while introducing significant security risks. These autonomous systems can plan, adapt, and interact with external environments, making them powerful but also susceptible to ...

Grok 3 AI Red Teaming

todayFebruary 18, 2025

  • 18258
  • 1
close

Research + LLM Security admin

Grok 3 Jailbreak and AI red Teaming

Grok 3 Jailbreak and AI Red Teaming In this article, we will demonstrate  how Grok 3 respond to different hacking  techniques including Jailbreaks and Prompt leaking attacks. Our initial study on AI Red Teaming different LLM Models using various approaches focused on LLM models released before the so-called “Reasoning Revolution”, ...

todayFebruary 17, 2025

  • 101
close

Secure AI Weekly admin

Towards Secure AI Week 6 – New AI Security Framework

Announcing the Databricks AI Security Framework 2.0 DataBricks, February 12, 2025 Databricks has unveiled the second edition of its AI Security Framework (DASF 2.0), a comprehensive guide designed to address the growing risks associated with AI deployments. The framework identifies 62 technical AI risks and introduces 64 mitigation controls, offering ...

todayFebruary 12, 2025

  • 94
close

Secure AI Weekly admin

Towards Secure AI Week 5 – Worldwide AI safety report

World-leading AI cyber security standard to protect digital economy and deliver Plan for Change Gov UK, January 31, 2025 The UK government has unveiled a pioneering cybersecurity standard aimed at protecting artificial intelligence (AI) systems from cyber threats, reinforcing the security of the digital economy. This initiative seeks to ensure ...

todayFebruary 5, 2025

  • 95
close

Secure AI Weekly admin

Towards Secure AI Week 4 – DeepSeek’s AI Security Failures

Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History Wiz, January 29, 2025 A recent security lapse in AI infrastructure has underscored the critical need for stronger protections in artificial intelligence systems. Wiz Research uncovered an unprotected ClickHouse database belonging to DeepSeek, a Chinese AI startup known ...

todayJanuary 31, 2025

  • 18840
close

Research + LLM Security admin

DeepSeek Jailbreak’s

Deepseek Jailbreak’s In this article, we will demonstrate how DeepSeek respond to different jailbreak techniques. Our initial study on AI Red Teaming different LLM Models using various aproaches focused on LLM models released before the so-called “Reasoning Revolution”, offering a baseline for security assessments before the emergence of advanced reasoning-based ...