AIResearch AIResearch
Back to articles
Science

AI Training Method Prevents Security Failures by Strengthening Adversarial Attacks

New adaptive technique shields AI models from unexpected performance drops, making them more robust and faster to train without sacrificing accuracy.

AI Research
November 14, 2025
3 min read
AI Training Method Prevents Security Failures by Strengthening Adversarial Attacks

Artificial intelligence systems, especially those used in security-critical applications like autonomous vehicles and facial recognition, often fail when faced with subtle, human-imperceptible changes to input data. These so-called adversarial attacks can trick models into making incorrect predictions, raising concerns about their reliability. A recent study challenges the long-held belief that this vulnerability stems from overfitting during training, instead identifying underfitting in the attack generation process as the primary culprit. This insight has led to a new method that not only prevents performance degradation but also speeds up training significantly, offering a more efficient path to robust AI.

The key finding from the research is that the robustness drop in adversarial training—where a model's ability to withstand attacks decreases after prolonged training—is primarily caused by underfitting in the perturbation generator, not overfitting as commonly assumed. Perturbations are small, crafted changes to input data designed to test and improve model resilience. The analysis shows that when these perturbations weaken over time, they fail to challenge the model adequately, leading to a collapse in performance. For instance, on the CIFAR-10 dataset using a Pre-ResNet18 model, adversarial accuracy under a PGD-20 attack dropped from 44.36% at the 20th epoch to 0% at the 30th epoch, highlighting the severity of this issue.

To investigate this, the researchers employed a systematic methodology centered on estimating the strength of perturbations. They defined a gap metric to measure how close generated perturbations are to the ideal worst-case attacks. By visualizing this gap over training epochs, they observed that for methods like FGSM (Fast Gradient Sign Method), the perturbation strength increases initially but then deteriorates to noise-like levels, coinciding with the robustness drop. This was validated through experiments where perturbations generated at critical epochs were replaced with random noise, resulting in similar performance declines, confirming that weak perturbations are to blame.

The results analysis, supported by figures in the paper, demonstrates that strengthening the perturbation generator mitigates the drop. For example, in Figure 1(b), Acc(FGSM, FGSM) and Acc(FGSM, Random) curves show that FGSM-generated perturbations lose effectiveness over time, while enhancements like FGSM+ (a variant that parameterizes the generator) alleviate this issue. The proposed solution, APART (Adaptive Adversarial Training), builds on this by factorizing input perturbations into layer-wise components and using learnable parameters to progressively strengthen the generator. In tests on CIFAR-10 with Pre-ResNet18, APART achieved 49.30% accuracy under PGD-20 attacks, outperforming PGD-10's 47.43%, and did so with about 4 times faster training, as shown in Figure 2.

In practical terms, this matters because it makes AI models more reliable and efficient for real-world use. Adversarial training is essential for applications where security is paramount, such as in medical imaging or financial systems, where even minor data manipulations could lead to critical errors. By addressing the root cause of robustness drops, APART enables models to maintain high performance without the computational overhead of traditional methods. This could lead to wider adoption of robust AI in industries that demand both accuracy and speed, reducing risks associated with adversarial threats.

However, the study acknowledges limitations, such as the trade-off between clean accuracy and robustness. As noted in the paper, improvements in adversarial performance often come at a slight cost to accuracy on unperturbed data. Additionally, the method's effectiveness across diverse architectures, like Transformers, remains unexplored, and further work is needed to explicitly regularize the generator gap to prevent underfitting in other contexts. These areas highlight opportunities for future research to refine and expand the approach.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn