AIResearch AIResearch
Back to articles
Science

AI Learns to Think Better by Focusing on Key Steps

A new method trains AI models to prioritize the most important reasoning steps, boosting accuracy on complex math and coding tasks without relying on guesswork or external guidance.

AI Research
November 14, 2025
3 min read
AI Learns to Think Better by Focusing on Key Steps

Artificial intelligence systems are increasingly tasked with solving complex problems that require step-by-step reasoning, but traditional training methods often treat every step equally, leading to inefficiencies and errors. A new study introduces VCORE, an optimization-based approach that reweights the importance of each reasoning token during training, significantly enhancing AI performance on challenging tasks like mathematical proofs and code generation. This advancement addresses a critical bottleneck in AI development, where misallocated training signals can undermine generalization and reliability.

The key finding is that VCORE consistently outperforms existing methods by dynamically adjusting token weights based on their utility in reducing the overall training objective. Unlike uniform weighting, which spreads supervision evenly across all tokens, VCORE identifies and emphasizes the most impactful steps in a reasoning chain. For instance, in evaluations on models like Qwen3 and LLaMA-3.1-8B-Instruct, VCORE achieved an average accuracy of 31.03% across benchmarks, surpassing alternatives such as Dynamic Fine-Tuning (28.97%) and standard supervised fine-tuning (28.35%). This improvement was particularly notable in out-of-domain settings, where VCORE maintained robust performance on tasks like R-Bench and SGPQA-1k, demonstrating better generalization.

Methodologically, VCORE formulates token reweighting as a constrained optimization problem, maximizing the first-order decrease in the training loss while keeping the weight distribution close to uniform to ensure stability. It employs a closed-form solution derived from stochastic gradient descent dynamics, avoiding reliance on heuristics or external guidance. A key innovation is the 'one-backward trick,' which efficiently estimates token utilities with minimal computational overhead—requiring only one additional forward pass per batch. This makes VCORE scalable and easy to integrate into standard training pipelines without architectural changes.

Results from the paper show that VCORE not only improves accuracy but also stabilizes training. For example, on the Olympiad benchmark, VCORE's accuracy increased from 32.70% to 65.0% as training data scaled from 4k to 32k samples, outperforming baselines that degraded with larger datasets. Component analyses revealed that variance-controlled scaling is essential; without it, sharp reweighting can amplify noise and hinder convergence. In ablation studies, VCORE maintained performance across a range of hyperparameters, with optimal results at a temperature setting of 1e-4, indicating robustness to parameter variations.

In practical terms, VCORE's implications are significant for real-world applications where AI must reason accurately, such as in educational tools, automated programming, and scientific research. By providing a more effective initialization for reinforcement learning, VCORE also enhances subsequent training phases, as shown in experiments where VCORE-initialized models achieved higher post-reinforcement learning scores on tasks like BigMath. This could lead to more reliable AI assistants capable of handling intricate, multi-step problems without excessive computational resources.

However, the study acknowledges limitations, including its focus on datasets like OpenMathReasoning and OpenCodeReasoning, which may not capture all reasoning scenarios. Potential failure modes, such as overemphasizing spurious tokens that leak answer information, were noted but not fully explored. Future research should investigate diverse datasets and integrate regularization techniques to mitigate these risks, ensuring VCORE's adaptability across broader contexts.

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn