AIResearch AIResearch
Back to articles
Data

AI Experts Learn Better When They See Clearer Data

A new method uses a special loss function to untangle messy data, allowing AI models to specialize more effectively and improve accuracy on complex tasks like image recognition.

AI Research
April 01, 2026
3 min read
AI Experts Learn Better When They See Clearer Data

A new approach to building AI models helps them avoid a common pitfall where multiple components learn the same thing, leading to wasted effort and reduced performance. Researchers from De La Salle University have enhanced a type of AI architecture called Mixture-of-Experts (MoE), which uses a group of specialized neural networks, or 'experts,' each handling different parts of a dataset under the guidance of a 'gating' network. The issue, known as 'expert collapse,' occurs when overlapping data boundaries cause experts to develop redundant representations, forcing the gating network into rigid and inefficient routing. By pre-processing data with a technique called Soft Nearest Neighbor Loss (SNNL), the team disentangles the input features, making it easier for experts to specialize uniquely, which boosts classification accuracy on challenging image datasets like FashionMNIST, CIFAR10, and CIFAR100.

The key finding is that SNNL resolves expert collapse by transforming raw input features into a latent space where similar data points cluster closely together. This disentanglement simplifies the gating network's task of assigning data to experts, allowing for more flexible routing. The researchers measured this using two new metrics: Expert Specialization Entropy, which quantifies routing flexibility, and Pairwise Embedding Similarity, which assesses how orthogonal the experts' learned weights are. On complex datasets, the SNNL-augmented MoE models showed statistically significant improvements in accuracy, with FashionMNIST increasing from 91.33% to 91.61% and CIFAR100 from 35.75% to 36.74%, as detailed in Table 1 of the paper. These gains come from experts learning distinct representations, reducing redundancy and enabling better collaboration.

Ology involves adding a feature extractor network optimized with SNNL before the MoE routing phase. This feature extractor, illustrated in Figure 1 of the paper, uses a convolutional neural network (CNN) with two blocks of layers to process raw input images into a structured latent representation. The SNNL function minimizes distances between class-similar data points in this representation, acting as a regularizer during training. The composite loss function combines cross-entropy for classification with SNNL, weighted by a factor alpha, to shape the feature space. All experiments were conducted on an RTX 3060 GPU using PyTorch Lightning, with models trained for 15,000 steps on benchmark datasets like MNIST, FashionMNIST, CIFAR10, and CIFAR100, ensuring reproducibility across multiple random seeds.

Analysis of reveals that SNNL promotes expert diversity and routing flexibility. For instance, on CIFAR100, the baseline model had a Pairwise Embedding Similarity of 0.20, indicating high redundancy among experts, but the SNNL model reduced this to 0.10, as shown in Figure 4. This orthogonality allows the gating network to adopt a more distributed routing strategy, with higher entropy values observed in complex datasets. UMAP visualizations in Figure 6 confirm that SNNL creates dense, homogeneous class clusters in the latent space, unlike the entangled boundaries in baseline models. However, on simple datasets like MNIST, where baseline accuracy was already near 99.4%, SNNL introduced over-regularization, slightly degrading performance, highlighting that the benefits are most pronounced in complex, entangled feature spaces.

Of this research extend to real-world applications where AI models handle messy, high-dimensional data, such as in medical imaging or autonomous driving. By preventing expert collapse, could lead to more efficient and accurate AI systems that better leverage specialized components. The paper notes that this approach adds no computational overhead during inference, making it practical for resource-constrained environments. Future work, as recommended, could explore dynamic tuning of SNNL parameters, scaling to larger models like Vision Transformers, and comparing SNNL with other contrastive learning techniques to further optimize specialization in ensemble AI architectures.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn