AIResearch AIResearch
Back to articles
AI

AI Learns to Pick Better Data for Faster Training

A new method helps large language models select the most useful training samples on the fly, cutting data needs by 95% while improving performance on complex tasks.

AI Research
April 02, 2026
4 min read
AI Learns to Pick Better Data for Faster Training

Training large language models like those behind chatbots and translation tools typically requires massive datasets, consuming enormous computational resources and time. However, not all data is equally valuable—some samples teach the model more effectively than others. Researchers from the University of Illinois Chicago have developed a new approach that allows AI systems to intelligently select and weight training data as it arrives, dramatically reducing the amount of data needed while maintaining or even improving model performance. This breakthrough addresses a critical bottleneck in AI development, where data efficiency directly translates to lower costs and faster innovation cycles.

The key finding from their study is that an optimizer-aware framework for online data selection can significantly enhance the fine-tuning of large language models. Unlike traditional s that pre-select data offline or use simple gradient similarity scores, this new approach dynamically evaluates sample utility based on the current state of the optimization process. In experiments, their achieved competitive or superior performance using only 5% of the available training data compared to full-data training and existing selection baselines. For instance, on the TyDiQA multilingual question-answering benchmark, their approach improved F1 scores by substantial margins over s like TracIn and LESS, demonstrating that smarter data curation leads to better model outcomes even with severe data constraints.

Ology centers on a two-stage Filter-then-Weight algorithm that first identifies geometrically useful candidate samples and then optimizes their coefficients to form a precise composite update. This process is framed as an optimizer-aware update-matching problem, where the goal is to shape the next training update toward the target task under the geometry induced by adaptive optimizers like Adam. To make this practical for large models, the researchers introduced a factorized outer-product gradient representation and optimized matrix computations for long-context data, reducing computational overhead while preserving essential information. The algorithm operates in an online setting, where data arrives sequentially, and selection decisions must be made on-the-fly without access to the full dataset, mimicking real-world streaming scenarios.

From extensive experiments on benchmarks like MMLU and TyDiQA show that this consistently outperforms existing online data selection baselines under the same data budget. For example, with the Llama-3.2-1B model, their approach achieved a TyDiQA F1 score of 48.86, compared to 47.50 for GREATS and 47.21 for LESS, as detailed in Table 2 of the paper. The performance gains are particularly pronounced at smaller data ratios, where selecting informative samples is crucial, as illustrated in Figure 1, which shows their maintaining higher F1 scores as training data usage increases up to the 5% budget. Ablation studies further confirm that optimizer-awareness and the two-stage reweighting are essential components, with s lacking these elements showing degraded performance or instability.

Of this research extend beyond academic benchmarks to practical applications in AI development. By enabling models to learn more effectively from less data, this approach can reduce the environmental and financial costs associated with training large AI systems, making advanced language models more accessible. It also opens doors for continual learning scenarios where models must adapt to new information over time without catastrophic forgetting. The paper's suggest that future AI training pipelines could incorporate similar dynamic data selection mechanisms to optimize resource usage and improve model alignment with specific tasks.

Despite its successes, has limitations that warrant further investigation. The framework assumes a linearized approximation for adaptive optimizers like Adam, which may not hold in all training scenarios, potentially affecting selection accuracy. Additionally, the current implementation relies on random projection for dimensionality reduction, and sensitivity analyses show that performance can vary with projection dimensions, as indicated in Table 5 where higher dimensions generally yield better . The paper also notes that extending this framework to store and reuse historical gradients in a memory buffer is an interesting future direction, which could further enhance data efficiency in long-term training processes.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn