Knowledge about artificial intelligence and its security needs to be constantly improved
Cloud Security Podcast by Google, June 6, 2022
Cloud Security Podcast by Google held a conversation with Google Research Scientist Nicholas Carlini, during which Nicholas spoke about his vision of the security of AI and answered the relevant questions. This is the second time the topic of Secure AI is covered on this podcast. The previous one was considered by Deepmind’s CISO. So the topic is definitely getting momentum.
Among other things in the podcast, the topic of a threat model for a large-scale AI system and assessed attacks were explored. Nicholas talked about possible types of attacks and the threats associated with training machine learning systems on large datasets. For instance, the data was taken from the Internet thus couldn’t be trusted. Also Nicholas focused on three main issues researchers are worried about, namely Poisoning, Evasion, and Extraction attacks.
During the conversation, Nicholas talked about the future of safe and secure AI, potential threats, and how the “opacity” of AI affected the actions of attackers and defenders.
Overall it’s highly recommended to listen.
Listen to the full podcast record via the link.
RSAC, June 7, 2022
Google, Microsoft and MITER met at RSA, the world’s largest security conference, to discuss and understand the ubiquity of Red Teaming AI systems.
Specialists from Google, Microsoft, and MITER have tried to figure out if Red Teaming AI systems are the future, or if they’re just a whim. During the conference, they have to answer a number of topical questions:
– Are Red Teaming AI systems necessary, and what are the weights for organizations?
– What is the easiest way to protect AI systems in enterprises and what steps need to be taken for this?
The fact that Secure AI topic was at the roundtable of the most popular Cybersecurity conference and brought experts from top industry players is definitely a sign that Securing AI is a must-have.
Meritalk, June 7, 2022
To wrap up, the interest on Secure AI, DARPA is continuing its Research towards Trustworthy AI
As part of its latest Assured Neuro Symbolic Learning and Reasoning (ANSR) program, the Defense Advanced Research Projects Agency (DARPA) is aiming for problem-solving proposals in artificial intelligence (AI) architectures and algorithms that integrate symbolic reasoning with data-driven machine learning. Thus, the Agency seeks to create trustworthy and reliable systems based on AI.
In a press release, DARPA ANSR Program Manager Sandeep Neema, said: “Motivating new thinking and approaches in this area will help ensure that autonomous systems operate safely and work properly”, “This will be an integral part of trust, which is the key to a successful acceptance of autonomy by the Ministry of Defense.”
What is the overall goal of the ANSR program? It is the refinement of hybrid AI algorithms and the elaboration of methods based on evidence that supports confident veracity judgments for these algorithms.
Even with recent improvements in machine learning algorithms and guaranteed technologies, high-level autonomy continues to be elusive. Based on the words of DARPA officials, the reason for this is twofold:
- On the one hand, data-driven machine learning flows robustness, interpretability, and transparency, while computational and data needs are unstable;
- On the other hand, standard approaches to the development of intelligent applications and autonomous systems based on knowledge representation and symbolic reasoning, can be guaranteed, but remain unstable to the uncertainties encountered in the real world.
Read more about DARPA ANSR Program in the article.
Subscribe for updates
Stay up to date with what is happening! Get a first look at news, noteworthy research and worst attacks on AI delivered right in your inbox.