AIResearch AIResearch
Back to articles
Coding

AI Learns Faster by Predicting Whole Text Segments at Once

New training method reduces errors in language models, speeding up learning without extra data—key for efficient AI in science and business.

AI Research
November 14, 2025
3 min read
AI Learns Faster by Predicting Whole Text Segments at Once

In the race to build smarter artificial intelligence, training large language models often hits a roadblock: inefficiency. These models, which power everything from chatbots to research tools, learn by predicting hidden words in text, but random guessing during training introduces noise that slows progress. A new study tackles this by restructuring how AI processes information, leading to faster and more reliable learning without requiring additional data. This breakthrough could cut development time and costs for companies and researchers relying on AI for complex tasks.

The key finding is that a novel masking strategy, called fully-explored masking, reduces variance in the training process. Variance refers to the fluctuations in learning signals that make optimization less stable. By minimizing this, the method enables language models to learn more efficiently from the same amount of data. In experiments, models trained with this approach consistently outperformed standard methods in tasks like text classification and relation extraction, showing improved generalization.

Methodology involved dividing text into non-overlapping segments and masking all tokens in one segment at a time during training. For example, if a sentence is split into four parts, the model predicts the words in one part while using the others as context. This contrasts with traditional methods that randomly mask individual words or spans, which can lead to inconsistent learning due to varying prediction difficulties. The researchers provided a theoretical analysis linking this approach to reduced gradient covariance, supported by empirical data showing that variance decreases as the Hamming distance between masks increases.

Results analysis from the paper demonstrates clear advantages. In continual pre-training scenarios using domains like computer science and news, the fully-explored method achieved higher scores on benchmarks. For instance, on the ACL-ARC dataset, it reached 76.24 with subword masking compared to 75.34 for the baseline, and similar gains were seen in other tasks like SciERC and HyperPartisan detection. In general pre-training from scratch, the method improved performance on 7 out of 8 GLUE benchmark tasks, with an average score increase from 82.5 to 83.6. Figure 4 in the paper illustrates that models using this strategy learn faster in early training steps, requiring fewer iterations to reach comparable accuracy.

Contextually, this matters because it addresses a core challenge in AI development: making training more sample-efficient. For businesses, it means AI systems can be adapted to specific domains like legal or medical text with less data and computation, reducing costs and time-to-market. In research, it allows for quicker experimentation and deployment of models in data-scarce environments, potentially accelerating discoveries in fields reliant on natural language processing.

Limitations noted in the study include the sensitivity to hyperparameters like the number of segments, though performance remains stable across variations. The method assumes fixed-length sequences and may not fully eliminate variance from other sources, such as data sampling. Further research is needed to explore its applicability to longer texts or other AI architectures beyond language models.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn