Towards Trusted AI Week 43 – AI TRISM is in the Gartner top Tech Trends 2024

Secure AI Weekly + Trusted AI Blog admin todayOctober 24, 2023 47

Background
share close

Enterprises struggle to address generative AI’s security implications

AINews, October 18, 2023

A recent study by the cloud-native network detection firm, ExtraHop, reveals a pressing concern in today’s digital age: organizations are finding it challenging to navigate the security implications of their employees using generative AI. The report, titled “The Generative AI Tipping Point,” delves deep into the complexities businesses face with the rising ubiquity of generative AI tools in daily operations. Alarmingly, while 73% of IT and security leaders confirmed their employees’ regular use of such tools, many felt ill-equipped to handle the associated security challenges.

Security concerns are multifaceted. On one hand, 40% of leaders feared AI tools delivering inaccurate or non-relevant outputs. Yet, this overshadowed serious security threats: only 36% were concerned about the potential exposure of sensitive customer or employee data, and a mere 25% were apprehensive about financial repercussions. Raja Mukerji of ExtraHop emphasized the potential of generative AI, stating, “With innovation balanced by robust protection, generative AI can indeed revolutionize industries in the near future.”

However, merely prohibiting the use of generative AI has proven ineffective. While 32% of participants indicated their organizations had bans in place, compliance was meager at 5%. This disparity underlines the need for more comprehensive approaches. The clamor for governmental oversight is evident, with 90% of respondents advocating for it, highlighting the gaps and dire need for bolstered security practices in the age of AI.

Gartner Top 10 Strategic Technology Trends for 2024

Gartner, October 16, 2023

Embracing AI’s potential and challenges is becoming increasingly crucial as technological advancements shape business and technical decisions for the coming years. The safety and security aspects of AI are not just innovative cornerstones but also fundamental in fostering trust. Key trends such as AI Trust, Risk, and Security Management (AI TRiSM) highlight the significance of creating reliable AI systems. In contrast, Continuous Threat Exposure Management (CTEM) underscores the continuous need for vigilance against potential threats.

Three central themes encompass these developments. The first theme underscores the importance of focused, secure, and future-oriented AI innovations. The second promotes tailored tech solutions and an inclusive environment for innovation that aligns with stakeholder needs. The final theme emphasizes the adaptive nature of AI, which includes catering to the growing influence of algorithm-driven customers and ensuring controlled access to the latest digital tools.

For businesses navigating this dynamic environment, it’s vital to contextually evaluate each AI trend and incorporate it into their strategic outlook. While these trends are distinct, they’re also interconnected, necessitating a comprehensive approach from businesses. By embracing these developments, organizations can bolster resilience, optimize data utilization, and accelerate digital evolution. In today’s rapidly changing tech landscape, the security and safety of AI are central to building lasting trust and value.

AI chatbots could help plan bioweapon attacks, report finds

The Guardian, October 17, 2023

Recent research by a US-based think tank, the Rand Corporation, has raised concerns over the potential misuse of artificial intelligence, specifically large language models (LLMs) like chatbots, in planning bioterrorist activities. The study, published on Monday, indicated that these AI models could provide guidance that might aid in strategizing a biological attack, though they stopped short of delivering explicit blueprints for weapon creation.

Historically, efforts to weaponize biological agents have often fallen short due to gaps in understanding the respective bacterium. For instance, the Japanese Aum Shinrikyo cult’s attempt in the 1990s to weaponize botulinum toxin was unsuccessful. The Rand Corporation’s report suggests that AI could quickly fill such knowledge voids, offering a dangerous accelerant to malicious intentions. Although the report refrained from naming the specific LLMs evaluated, it was shared that these models were accessed via an application programming interface (API). In various test scenarios set by the researchers, the AI models identified potential bioweapon agents, evaluated their potential impact, and even strategized cover stories for seemingly legitimate acquisitions of hazardous bacteria.

Given these findings, the report highlights the crucial necessity for rigorous AI model testing and stronger safety restrictions. The preliminary data insinuates that while some of the information LLMs provided might already be accessible online, the ease of obtaining it via AI presents a unique threat. As AI continues to advance and play an increasingly integral role in various sectors, its potential security and safety implications, such as the ones outlined in the Rand Corporation’s report, will be discussed at an upcoming global AI safety summit in the UK. This underscores the global community’s awareness and commitment to ensuring that AI’s power serves to benefit humanity, rather than endanger it.

Microsoft-affiliated research finds flaws in GPT-4

TechCrunch, October 17, 2023

A recent investigation affiliated with Microsoft has spotlighted the challenges and risks posed by the heightened precision of large language models (LLMs), specifically OpenAI’s GPT-4 and its predecessor, GPT-3.5. This research underscores a critical dilemma: GPT-4’s enhanced comprehension enables it to follow instructions meticulously, but this exactness can also be its downfall if manipulated malevolently. The study reveals that GPT-4 is susceptible to producing biased or harmful outputs when prodded with “jailbreaking” prompts—carefully crafted cues designed to circumvent its built-in safety mechanisms.

Microsoft’s participation in this research, especially when GPT-4 powers its Bing Chat chatbot, may appear counterintuitive. Yet, a nuanced reading suggests this might be a proactive attempt to identify and subsequently rectify potential vulnerabilities. Indications from the study hint that proactive measures might have been initiated even before its publication. This research serves as a sobering reminder: achieving absolute reliability and safety in LLMs is an ongoing challenge, even for tech behemoths.

One alarming revelation was GPT-4’s interaction with certain prompts. Maliciously designed prompts could lead the model down a path of generating inappropriate or prejudiced content, demonstrating underlying biases. Furthermore, GPT-4 showed a concerning propensity to unintentionally divulge sensitive data. To foster a collaborative approach towards mitigating such risks, the researchers have made their testing methodologies public on GitHub, inviting the global scientific community to contribute towards refining and safeguarding AI models.

Multi-modal prompt injection image attacks against GPT-4V

Simon Willison’s blog, October 14, 2023

OpenAI’s GPT-4V brings a groundbreaking feature to the fore: the ability to process and respond to images. Demonstrated by its apt description of an image from the “50th Annual Half Moon Bay Pumpkin Weigh-Off,” the model’s precision in extracting context from visuals is undeniably impressive. Such advancements highlight AI’s versatility and potential in diverse applications. However, with innovation often comes unforeseen challenges, and in this case, it’s the increased susceptibility to prompt injection attacks via images.

A series of experiments reveal GPT-4V’s vulnerability to visual prompt injections. In one instance, the model was easily misled by textual instructions within an image to “Say hello,” diverting it from its primary task of describing the image. More alarmingly, another test successfully executed an exfiltration attack. The AI was coaxed, using embedded instructions in an image, to encode a prior conversation into a URL and send it to an external server, raising significant data privacy concerns. Such experiments underscore the potential risks of embedding deceptive commands in images, be it overtly or subtly, like almost invisible off-white text on a white canvas.

This issue transcends GPT-4V’s image processing feature. The core challenge lies in the inherent nature of AI models to trustingly respond to the inputs they’re given, making them prime targets for prompt injections. While AI’s adaptability is its strength, discerning between legitimate and malicious instructions remains an intricate puzzle. As AI technologies continue to advance and integrate deeper into our daily lives, the need for robust security measures and a thorough understanding of their vulnerabilities becomes even more pressing.

 

Subscribe for updates

Stay up to date with what is happening! Get a first look at news, noteworthy research and worst attacks on AI delivered right in your inbox.

    Written by: admin

    Rate it
    Previous post