Analysis of the vulnerability of machine learning regression models to adversarial attacks using data from 5G wireless networks
By: Leonid Legashev, Artur Zhigalov, Denis Parfenov
Potential Business Impact:
Finds fake data that tricks computers.
This article describes the process of creating a script and conducting an analytical study of a dataset using the DeepMIMO emulator. An advertorial attack was carried out using the FGSM method to maximize the gradient. A comparison is made of the effectiveness of binary classifiers in the task of detecting distorted data. The dynamics of changes in the quality indicators of the regression model were analyzed in conditions without adversarial attacks, during an adversarial attack and when the distorted data was isolated. It is shown that an adversarial FGSM attack with gradient maximization leads to an increase in the value of the MSE metric by 33% and a decrease in the R2 indicator by 10% on average. The LightGBM binary classifier effectively identifies data with adversarial anomalies with 98% accuracy. Regression machine learning models are susceptible to adversarial attacks, but rapid analysis of network traffic and data transmitted over the network makes it possible to identify malicious activity
Similar Papers
Evaluating the Robustness of Adversarial Defenses in Malware Detection Systems
Cryptography and Security
Makes phone virus checkers easier to trick.
Evaluating the Impact of Adversarial Attacks on Traffic Sign Classification using the LISA Dataset
CV and Pattern Recognition
Makes self-driving cars see traffic signs better.
Adversarial Robustness Analysis of Vision-Language Models in Medical Image Segmentation
CV and Pattern Recognition
Makes AI better at spotting problems in medical pictures.