admin

360 Results / Page 11 of 40

todaySeptember 4, 2023

  • 152
close

Trusted AI Blog + LLM Security admin

LLM Security and Prompt Engineering Digest: Top August events, guides, incidents, VC reviews and research papers

Welcome to a brief exploration into the fascinating world of AI security—a realm where innovation and danger intertwine like DNA strands. Dive in to learn how red teaming tests AI vulnerabilities, what Google recommends for AI security, the unforeseen risks of AI in everyday applications, and academic approaches to the ...

todaySeptember 1, 2023

  • 71
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 35 – The Achilles’ Heel of AI

Tricks for making AI chatbots break rules are freely available online NewScientist, August 21, 2023 Artificial intelligence chatbots like ChatGPT have become essential tools for various online activities, but their security loopholes present an emerging concern. Manipulative text prompts, often referred to as “jailbreak prompts,” can mislead these AI systems ...

todayAugust 25, 2023

  • 146
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 34 – Defcon AI Red Teaming wrap-ups and the Quest for AI Security

Don’t expect quick fixes in ‘red-teaming’ of AI models. Security was an afterthought APNews, August 14, 2023 The recent DefCon hacker conference in Las Vegas served as a stark reminder of the pressing concerns around AI safety and security. The event saw 2,200 participants rigorously testing eight advanced language models, ...

todayAugust 7, 2023

  • 43
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 32 – Navigating the Future of Cyber Resilience

The generative A.I. battle between companies and hackers is starting CNBC, August 2, 2023 Last month, tech titans like Amazon, Google, Meta, and Microsoft collaborated with President Joe Biden, emphasizing their commitment to ensure that AI technologies undergo rigorous safety checks before public deployment. The primary concern is the role ...

todayAugust 3, 2023

  • 125
close

Secure AI Weekly + Trusted AI Blog admin

Towards Trusted AI Week 31 – New LLM Jailbreak, Plugin hacks and more

ChatGPT Has a Plugin Problem Wired, July 25, 2023 Over the past eight months, OpenAI’s ChatGPT has dazzled millions with its ability to produce lifelike text, from stories to code. However, the development and rapid proliferation of plugins to extend ChatGPT’s capabilities have raised serious security concerns. The introduction of ...

todayAugust 1, 2023

  • 99
close

Adversarial ML admin

Secure AI Research papers: Reviewing Strategic Offenses and Defenses in AI Models

This digest reviews four pivotal research papers that shed light on diverse dimensions of AI, from exploring vulnerabilities in Natural Language Inference (NLI) models and Generative AI to investigating adversarial attacks and defenses on 3D Point Cloud Classification, and unveiling the potential misuse of multi-modal LLMs.  Each study underlines the ...