Trusted AI Blog

283 Results / Page 4 of 32

todayDecember 6, 2023

  • 85
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 48 – Multiple OpenAI Security Flaws

OpenAI’s Custom Chatbots Are Leaking Their Secrets Wired, November 29, 2023 The rise of customizable AI chatbots, like OpenAI’s GPTs, has introduced a new era of convenience in creating personalized AI tools. However, this advancement brings with it significant security challenges, as highlighted by Alex Polyakov, CEO of Adversa AI. ...

todayDecember 1, 2023

  • 57
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 47 – UK Guides for secure AI development

AIs can trick each other into doing things they aren’t supposed to New Scientist, November 24, 2023 Recent developments in artificial intelligence (AI) have raised significant security concerns. Notably, AI models, which are generally programmed to reject harmful or illegal requests, have demonstrated a concerning ability to persuade each other ...

todayNovember 22, 2023

  • 74
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 46 – GPT’s Security Issues and OpenAI Drama

Top VC Firms Sign Voluntary Commitments for Startups to Build AI Responsibly Bloomberg, November 14, 2023 In a landmark initiative for the AI industry, over 35 leading venture capital firms, such as General Catalyst, Felicis Ventures, Bain Capital, IVP, Insight Partners, and Lux Capital, have committed to promoting responsible AI ...

todayNovember 16, 2023

  • 228
close

Adversarial ML + Trusted AI Blog admin

Secure AI Research Papers: Jailbreaks, AutoDAN, Attacks on VLM and more

Researchers explore the vulnerabilities that lie within the complex web of algorithms, and the need for a shield that can protect against unseen but not unfelt threats.   These papers published in October 2023 collectively study AI’s vulnerability, from the simplicity of human-crafted deceptions to the complexity of multilingual and visual ...

todayNovember 15, 2023

  • 76
close

Secure AI Weekly + Trusted AI Blog admin

Towards Secure AI Week 45 – LLM hacking LLM and new Google SAIF

Google’s Secure AI Framework  (SAIF) Google Google’s Secure AI Framework (SAIF) is a blueprint for securing AI and machine learning (ML) models, designed to be secure-by-default. It addresses concerns that are top of mind for security professionals, such as risk management, security, and privacy, ensuring that AI systems are safely ...

todayNovember 8, 2023

  • 146
close

Trusted AI Blog + LLM Security admin

LLM Security Digest: Best October’s Activities And Prompt Engineering Tricks

This digest of October 2023 encapsulates the most influential findings and discussions on LLM Security and a bit of Prompt Engineering. Subscribe for the latest LLM Security news: Jailbreaks, Attacks, CISO guides, VC Reviews and more   LLM Security  Best practical LLM Attacks: Multi-modal prompt injection image attacks against GPT-4V ...

todayNovember 6, 2023

  • 40
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 45 – US executive order and UK Bletchley Declaration on AI

FACT SHEET: President Biden Issues Executive Order on Safe, Secure, and Trustworthy Artificial Intelligence The White House, October 30, 2023 President Joe Biden has issued a decisive Executive Order to position the United States as a leader in the realm of artificial intelligence (AI), ensuring the nation captures the technology’s ...

todayNovember 2, 2023

  • 46
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 44 – Takeaways from Securing AI Summit

16 Takeaways from the 2023 Securing AI Summit in San Francisco BGV, October 24, 2023 The rapid integration of Generative AI (GenAI) technologies like ChatGPT into various industries is a double-edged sword. While these advancements promise significant productivity boosts, they also bring forth heightened concerns regarding cybersecurity. Just as we ...