Cybersecurity teams face a constant : intrusion detection systems (IDS) are often calibrated to recognize known attacks but struggle with new, unknown threats. This leaves networks vulnerable to zero-day attacks that exploit unseen patterns. A new AI framework offers a solution by generating realistic fake network traffic that enhances IDS training, improving detection of both familiar and novel attacks without requiring additional real data.
The researchers developed GMA-SAWGAN-GP, a generative model that creates synthetic network flow records mimicking real traffic. This approach addresses a key limitation in existing s: network data contains both discrete features (like protocol type) and continuous features (like packet size), which are hard to model together. By using Gumbel-Softmax regularization for discrete fields and a self-attention mechanism to capture dependencies between features, the generator produces high-quality synthetic samples that preserve the complex structure of real network traffic. An autoencoder acts as a manifold regularizer to keep generated data close to the real distribution, and an adaptive gating network balances different loss functions during training to prevent mode collapse and improve stability.
Extensive experiments on three public datasets—NSL-KDD, UNSW-NB15, and CICIDS2017—demonstrated the framework's effectiveness. Five different IDS models, including CNN, LSTM, and hybrid architectures, were trained on datasets augmented with synthetic samples. For binary classification (normal vs. abnormal traffic), accuracy improvements averaged 5.3% across all models and datasets. On UNSW-NB15, for example, a CNN-BiLSTM model saw a 9.3% accuracy boost, from 80.7% to 90.0%. In multi-classification tests, where specific attack types are identified, accuracy increased by an average of 2.2%, with the CNN-BiLSTM model improving by 6.5% on UNSW-NB15. The synthetic data helped models better distinguish between attack classes, particularly for minority attacks like Shellcode and Worms.
Generalization to unknown attacks was evaluated using the Leave-One-Attack-type-Out (LOAO) , where one attack type is excluded during training and reserved for testing. Models trained on augmented data showed higher Area Under the Receiver Operating Characteristic (AUROC) and True Positive Rate at a 5% False Positive Rate (TPR@5%FPR). For instance, on NSL-KDD with Probe as the unknown attack, a CNN model's AUROC increased from 84.4% to 85.6%, and on CICIDS2017 with PortScan unknown, an LSTM model's AUROC rose from 78.2% to 84.2%. Overall, AUROC for unknown attacks improved by 3.9% on average, and TPR@5%FPR increased by 4.8%, indicating better detection at low false alarm rates.
Are significant for real-world cybersecurity. By generating diverse synthetic traffic, this allows IDS models to explore a broader range of attack patterns, making them more robust to novel threats without costly data collection. It's particularly useful for rare attack types, where real samples are scarce. However, the study has limitations: it relies on historical datasets that may not fully reflect modern network environments, and it excludes newer IDS architectures like Transformers. Additionally, performance gains can vary with extremely sparse unknown attacks, and requires careful tuning to avoid distribution mismatches. Despite this, the framework represents a step toward more adaptive and resilient intrusion detection in an evolving threat landscape.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn