AIResearch AIResearch
Back to articles
Coding

New Super-Metric Stabilizes Synthetic Data Evaluation for Android Malware Detection

Researchers integrate a composite scoring system into the MalDataGen framework, addressing fragmentation and inconsistency in synthetic data quality assessment.

AI Research
March 26, 2026
4 min read
New Super-Metric Stabilizes Synthetic Data Evaluation for Android Malware Detection

In the high-stakes arena of Android malware detection, machine learning models are only as good as the data they're trained on. Yet, acquiring real, comprehensive, and high-quality datasets remains a persistent bottleneck, often hampering model performance and generalization. Synthetic data generation has emerged as a promising strategy to mitigate this scarcity, artificially creating datasets that mimic real-world patterns. However, as researchers from Horizon IA Labs and the Federal University of Pampa highlight in a new study, evaluating the quality of this synthetic data is fraught with instability and a lack of standardization. With over 65 distinct fidelity metrics reported in the literature, applied independently and inconsistently, the field suffers from fragmentation that hinders model-to-model comparison and experimental reproducibility. This ological chaos complicates the integrated interpretation of data quality, posing a significant barrier to deploying synthetic data in critical cybersecurity applications where reliability is paramount.

The researchers set out to transform this landscape by enhancing the MalDataGen framework, an existing modular, open-source platform designed for synthetic tabular data generation in Android malware contexts. Their central innovation is the integration of a Super-Metric, a composite scoring system that aggregates eight individual metrics across four fundamental dimensions: Distance, Correlation/Association, Feature Similarity, and Multivariate Distribution. This Super-Metric produces a single weighted score, aiming to reduce the variability and inconsistency observed when relying on isolated metrics. By incorporating this into MalDataGen, the framework evolves from a mere generation tool into a complete ecosystem for multidimensional generation and evaluation. The generative layer of MalDataGen itself is comprehensive, including four groups of models: Adversarial Models (GANs like classical GAN, WGAN, and WGAN-GP), Autoencoders (standard, VAE, and quantized VAE), Diffusion Models (Denoising Diffusion and Latent Diffusion), and Statistical/third-party models (SMOTE and SDV library models such as CTGAN, TVAE, and Copula). This diversity enables rigorous benchmarking under identical conditions, a critical step for meaningful comparison.

To assess the Super-Metric's effectiveness, the team conducted extensive experiments involving ten generative models and five balanced Android malware datasets. For each combination, they computed traditional fidelity metrics alongside the proposed Super-Metric, comparing these against utility metrics—specifically recall and F1-score—obtained from classifiers trained on synthetic data and evaluated exclusively on real data. The Super-Metric was computed separately for each dataset, with weights adjusted to minimize the gap between recall and F1-score values, ensuring the score reflects practical utility in classification tasks. The analysis focused on three key properties: consistency (maintaining the same correlation sign), stability (low variance across models), and robustness (behavior independent of generative architecture). Visualizations included heatmaps showing average correlations and boxplots representing distribution variations, as referenced in Figures 2 and 3 of the paper.

Revealed a stark contrast between traditional metrics and the new Super-Metric. Traditional metrics exhibited highly unstable behavior, alternating between positive and negative correlations and displaying large variation across different generative models. This inconsistency indicates that none of these isolated metrics can serve as a universal fidelity indicator. In contrast, the Super-Metric demonstrated greater stability, a consistent correlation sign, and better alignment with recall and F1-score. Even when it did not achieve the highest absolute correlation in every case, it stood out as the most stable metric across generators, with its weighted aggregation effectively reducing noise and mitigating limitations present in metrics evaluated in isolation. The heatmaps and boxplots illustrated this advantage, highlighting the Super-Metric's superior performance in heterogeneous generation scenarios, where it provided a more reliable predictor of synthetic data's real-world impact on classifier performance.

This work has significant for the field of cybersecurity and synthetic data evaluation. By providing a more robust, consistent, and contextualized evaluation , the Super-Metric integrated into MalDataGen addresses a critical gap in quality assessment, making synthetic data more viable for high-risk applications like malware detection. The framework's evolution into a complete benchmarking platform enables reproducible and ologically sound experiments, fostering greater trust in synthetic data's utility. Moreover, the approach's generalizability suggests potential applications beyond Android malware, possibly extending to other domains where tabular data synthesis is crucial, such as finance or healthcare, though this would require further validation.

Despite its advancements, the study acknowledges limitations and outlines directions for future work. The Super-Metric, while more stable, may still benefit from advanced optimization techniques like evolutionary algorithms, meta-learning, or nonlinear combination strategies to enhance its accuracy further. Expanding the approach to other domains and introducing interpretability mechanisms to understand each dimension's contribution are promising next steps. Additionally, integrating the Super-Metric into MLOps pipelines could enable continuous monitoring of synthetic data quality in real production environments, reinforcing its role as a central component within the MalDataGen ecosystem. These developments could help solidify synthetic data as a reliable tool in the ever-evolving battle against cyber threats.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn