Aiding medics and law enforcement

Adversarial ML admin todayDecember 31, 2019 255

Background
share close

This is Adversa’s monthly digest of the studies in the field of Machine Learning and Artificial Intelligence. From December 2019 we choose to cover the studies that look at AI applications in our healthcare and public safety systems.


Robust, Extensible, and Fast: Teamed Classifiers for Vehicle Tracking and Vehicle Re-ID in Multi-Camera Networks 

What makes vehicle identification and tracking so complex? Traffic camera footage differs in resolution, scale, and orientation. In general, cars are similar. Yet the same car can look differently depending on perspective and lighting. Suprem et al. suggest turning to teamed classifiers to address these problems. First, a team of functions performs coarse clustering of data based on features such as color or model of a car. Then each cluster is analyzed separately and simultaneously. The initial division into subsets allows for higher accuracy overall. The fact that the classifier is built from smaller models allows it to operate in near real-time. And since the functions are trained independently, the system can be continuously updated with the release of new vehicles. 


Universal Adversarial Perturbations for CNN Classifiers in EEG-Based BCIs 

Brain-computer interfaces (BCIs) enable people to control wheelchairs and exoskeletons using brain signals recorded by electroencephalograms (EGG). To automate the decoding of those signals scientists suggest using convolutional neural networks (CNNs). However, CNNs are vulnerable to adversarial attacks. Liu, Zhang, and Wu prove that using a highly flexible total loss minimization (TLM) approach it is possible to design universal adversarial perturbations that can be added to any EGG in real-time. This would allow attackers to affect the diagnosis of a disabled person or take control of the user’s wheelchair or exoskeleton, putting them in danger.


Segmentations-Leak: Membership Inference Attacks and Defenses in Semantic Image Segmentation 

The availability of large data sets set off the surge in AI development. However, they can be a vulnerability in themselves. If a data set includes private information, models trained using it are susceptible to membership inference attacks, where an adversary can infer whether or not a particular piece of sensitive data was part of the training dataset by looking at the outputs of the model. To protect users He et al. improved upon this class of attacks and then came up with solutions to defer them. Namely, they recommend adding Gaussian noises to data and applying DPSGD to train a model.


CALL TO ACTION

Written by: admin

Rate it

Previous post