Top funny AI incidents of all time
Artificial intelligence is far from perfect – this is no secret to anyone. Despite the fact that some of his mistakes and attacks can lead to serious consequences, some are ...
Secure AI Weekly admin todayFebruary 28, 2022 64
ML models not as secure as it might seem at first glance
Vice, February 22, 2022
Faces created by artificial intelligence become so believable that they blur the line between real photos and fakes, even people sometimes find it difficult to distinguish one from the other
Recently, a study by researchers Sophie J. Nightingale at Lancaster University in the UK and Hany Farid at the University of California, Berkeley was published, according to which people are very bad at distinguishing artificially generated images from real photographs.
The researchers conducted three experiments to determine how easy it is for people to tell deepfakes from real photos. In the first test, 315 participants, without prior preparation, tried to determine whether they were real or generated by looking at the photo. People were able to correctly identify the photo 48.2 percent of the time. In the second experiment, 219 other participants received comments as they passed the test; based on this, people more easily understood the parameters that are not characteristic of artificially generated images. This improved the statistics to 59 percent.
According to the study, it became clear that artificially generated images provide many new opportunities for scammers, since artificial photos do not arouse suspicion in many people.
The Register, February 25, 2022
The fact that the use of deep neural networks for computer vision tasks such as face recognition, medical imaging, object detection and autonomous driving is becoming more common may attract the attention of cybercriminals.
The application of DNNs is expected to grow rapidly in the coming years. Analysts at Emergen Research report that the global market for DNN technology will grow from $1.26 billion in 2019 to $5.98 billion by 2027, especially in industries such as healthcare, banking, financial services and insurance. Such fast-spreading attractions always attract the attention of visitors. They, in turn, can attack the learning process of AI models to inject hidden features or triggers into DNNs. It turns out a kind of Trojan horse for machine learning. Such a Trojan can change the behavior of the attackers to the desired one, which can have bad consequences. For example, at the request of the intruder, the system can quickly develop in people or objects. This, in turn, can lead to serious consequences – for example, in the case of autonomous cars that read traffic signs.
In their study, the experts propose two new defenses – variational input filtering (VIF) and adversarial input filtering (AIF). Both methods are designed to learn a filter that can detect all Trojan filters in the model’s input. The methods were applied to images and their classification.
Business Insider, February 25, 2022
The founder, whose company went bankrupt due to a Facebook algorithm tweak in 2018, commented on the Twitter incident in more detail this week.
Joe Spieser started LittleThings.com in 2014 as a women-centric digital media site. It focused on inspirational content such as animal videos, recipes, and other uplifting stories. The project was doing well until the fatal algorithm change. The fact is that the project’s 20 million social media followers were largely created from Facebook’s huge user base. On top of that, Facebook even used LittleThings as an example of how to build a successful media company at one of their annual conferences.
The problems started when traffic to the company’s pages was limited after Facebook changed its algorithm to promote posts that it thought people would interact with the most. For example, it could be messages from friends and family that the company believes could keep users on the platform longer. However, the algorithm has also begun promoting violent, false, and divisive content.
Stay up to date with what is happening! Get a first look at news, noteworthy research and worst attacks on AI delivered right in your inbox.
Written by: admin
Articles admin
Artificial intelligence is far from perfect – this is no secret to anyone. Despite the fact that some of his mistakes and attacks can lead to serious consequences, some are ...
Adversa AI, Trustworthy AI Research & Advisory