Performance of Machine Learning Classifiers for Anomaly Detection in Cyber Security Applications
By: Markus Haug, Gissel Velarde
Potential Business Impact:
Finds fake credit card charges better.
This work empirically evaluates machine learning models on two imbalanced public datasets (KDDCUP99 and Credit Card Fraud 2013). The method includes data preparation, model training, and evaluation, using an 80/20 (train/test) split. Models tested include eXtreme Gradient Boosting (XGB), Multi Layer Perceptron (MLP), Generative Adversarial Network (GAN), Variational Autoencoder (VAE), and Multiple-Objective Generative Adversarial Active Learning (MO-GAAL), with XGB and MLP further combined with Random-Over-Sampling (ROS) and Self-Paced-Ensemble (SPE). Evaluation involves 5-fold cross-validation and imputation techniques (mean, median, and IterativeImputer) with 10, 20, 30, and 50 % missing data. Findings show XGB and MLP outperform generative models. IterativeImputer results are comparable to mean and median, but not recommended for large datasets due to increased complexity and execution time. The code used is publicly available on GitHub (github.com/markushaug/acr-25).
Similar Papers
Cyber Security Data Science: Machine Learning Methods and their Performance on Imbalanced Datasets
Machine Learning (CS)
Finds computer threats faster by trying different tricks.
Malware Classification from Memory Dumps Using Machine Learning, Transformers, and Large Language Models
Machine Learning (CS)
Finds bad computer programs faster and better.
Enhancing IoT Cyber Attack Detection in the Presence of Highly Imbalanced Data
Machine Learning (CS)
Finds hidden internet dangers in busy networks.