admin

371 Results / Page 9 of 42

todayDecember 14, 2023

  • 107
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 49 – Multiple Loopholes in LLM… Again

LLMs Open to Manipulation Using Doctored Images, Audio Dark Reading, December 6, 2023 The rapid advancement of artificial intelligence (AI), especially in large language models (LLMs) like ChatGPT, has brought forward pressing concerns about their security and safety. A recent study highlights a new type of cyber threat, where attackers ...

todayDecember 8, 2023

  • 565
close

Trusted AI Blog + LLM Security admin

LLM Security Digest: Hacking LLM, Top LLM Attacks, VC Initiatives, LLM Incidents and Research papers in November 

This digest of November 2023 keeps the essential findings and discussions on LLM Security. From Hacking LLM using the intriguing ‘Prompt-visual injections’ to the complex challenges in securing systems like Google Bard, we cover the most crucial updates.   Subscribe for the latest LLM Security and Hacking LLM news: Jailbreaks, ...

todayDecember 6, 2023

  • 122
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 48 – Multiple OpenAI Security Flaws

OpenAI’s Custom Chatbots Are Leaking Their Secrets Wired, November 29, 2023 The rise of customizable AI chatbots, like OpenAI’s GPTs, has introduced a new era of convenience in creating personalized AI tools. However, this advancement brings with it significant security challenges, as highlighted by Alex Polyakov, CEO of Adversa AI. ...

todayDecember 1, 2023

  • 93
close

Trusted AI Blog + Secure AI Weekly admin

Towards Secure AI Week 47 – UK Guides for secure AI development

AIs can trick each other into doing things they aren’t supposed to New Scientist, November 24, 2023 Recent developments in artificial intelligence (AI) have raised significant security concerns. Notably, AI models, which are generally programmed to reject harmful or illegal requests, have demonstrated a concerning ability to persuade each other ...

todayNovember 22, 2023

  • 102
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 46 – GPT’s Security Issues and OpenAI Drama

Top VC Firms Sign Voluntary Commitments for Startups to Build AI Responsibly Bloomberg, November 14, 2023 In a landmark initiative for the AI industry, over 35 leading venture capital firms, such as General Catalyst, Felicis Ventures, Bain Capital, IVP, Insight Partners, and Lux Capital, have committed to promoting responsible AI ...

todayNovember 16, 2023

  • 329
close

Trusted AI Blog + Adversarial ML admin

Secure AI Research Papers: Jailbreaks, AutoDAN, Attacks on VLM and more

Researchers explore the vulnerabilities that lie within the complex web of algorithms, and the need for a shield that can protect against unseen but not unfelt threats.   These papers published in October 2023 collectively study AI’s vulnerability, from the simplicity of human-crafted deceptions to the complexity of multilingual and visual ...

todayNovember 15, 2023

  • 118
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 45 – LLM hacking LLM and new Google SAIF

Google’s Secure AI Framework  (SAIF) Google Google’s Secure AI Framework (SAIF) is a blueprint for securing AI and machine learning (ML) models, designed to be secure-by-default. It addresses concerns that are top of mind for security professionals, such as risk management, security, and privacy, ensuring that AI systems are safely ...

todayNovember 8, 2023

  • 196
close

Trusted AI Blog + LLM Security admin

LLM Security Digest: Best October’s Activities And Prompt Engineering Tricks

This digest of October 2023 encapsulates the most influential findings and discussions on LLM Security and a bit of Prompt Engineering. Subscribe for the latest LLM Security news: Jailbreaks, Attacks, CISO guides, VC Reviews and more   LLM Security  Best practical LLM Attacks: Multi-modal prompt injection image attacks against GPT-4V ...