AIResearch AIResearch
Back to articles
Coding

New AI Training Method Boosts Reasoning by Fine-Tuning Turn-by-Turn Rewards

In the fast-evolving world of artificial intelligence, training large language models (LLMs) to reason with tools like code interpreters has been a persistent , often hitting performance plateaus with…

AI Research
November 20, 2025
4 min read
New AI Training Method Boosts Reasoning by Fine-Tuning Turn-by-Turn Rewards

In the fast-evolving world of artificial intelligence, training large language models (LLMs) to reason with tools like code interpreters has been a persistent , often hitting performance plateaus with existing reinforcement learning s. A groundbreaking study by researchers from the University of Illinois Urbana-Champaign, AWS AI Labs, and Meta introduces Group Turn Policy Optimization (GTPO), a novel algorithm designed specifically for multi-turn Tool-Integrated Reasoning (TIR), where models iteratively generate code, execute it, and refine their reasoning over multiple steps. This approach addresses critical flaws in current techniques like Group Relative Policy Optimization (GRPO), which struggles with coarse, trajectory-level rewards that fail to provide sufficient learning signals for complex, multi-step tasks. By focusing on fine-grained, turn-by-turn feedback, GTPO promises to push the boundaries of AI reasoning, making models more adept at handling real-world mathematical and logical problems with greater precision and efficiency.

Current reinforcement learning s for TIR, such as GRPO, assign a single reward to an entire reasoning trajectory, leading to noisy and ineffective training signals in multi-turn scenarios. The authors highlight that this sequence-level approach ignores the dynamic nature of TIR, where each turn—comprising text generation, tool invocation, and execution feedback—can drastically alter the model's reasoning path. For instance, in tasks like solving mathematical problems, a model might generate incorrect code in one turn but correct it in the next, yet GRPO's all-or-nothing reward system fails to capture these nuances. This in training stagnation, where models stop improving despite continued iterations, as evidenced by empirical observations in the paper. The limitations are compounded by sparse binary rewards based solely on final answer accuracy, which overlook partially correct steps and hinder the model's ability to learn from intermediate successes or failures.

To overcome these issues, GTPO introduces three key innovations: turn-level reward assignment, return-based advantage estimation, and self-supervised reward shaping. Turn-level rewards provide individualized feedback for each reasoning turn, such as penalizing format errors or rewarding correct tool usage, rather than lumping everything into a single outcome. Return-based advantages incorporate a discount factor to account for the temporal sequence of turns, ensuring that rewards from later steps are weighted appropriately relative to earlier ones. Meanwhile, self-supervised reward shaping densifies sparse binary rewards by using code similarity scores—comparing generated code in incorrect trajectories against correct ones—to assign partial rewards, thus leveraging valuable learning signals even from failed attempts. In evaluations on benchmarks like AIME 2024 and MATH 500, GTPO achieved an average 3.0% improvement over GRPO, with models showing higher accuracy and more stable training curves, as detailed in the study's comprehensive experiments.

Of GTPO extend beyond academic benchmarks, potentially revolutionizing how AI systems handle complex reasoning in fields like software engineering, data analysis, and scientific research. By enabling more efficient and nuanced training, GTPO could lead to LLMs that better integrate external tools, reducing errors and improving reliability in real-world applications. The authors note that this approach aligns with broader trends in AI toward multi-turn interactions, where models must adapt dynamically based on feedback, much like human problem-solving. However, they caution that current experiments are limited to 7B-parameter models due to computational constraints, and scaling to larger models or diverse domains remains an area for future work. Despite this, 's simplicity and effectiveness suggest it could be widely adopted, fostering advancements in AI reasoning that were previously hampered by training inefficiencies.

Despite its promising , GTPO has limitations, including its focus on mathematical reasoning tasks and the resource-intensive nature of RL training, which restricted testing to smaller models. The authors acknowledge that broader applicability to domains like software development or general multi-turn scenarios requires further validation. Additionally, the choice of hyperparameters, such as the discount factor, was found to be crucial, with optimal performance at γ=0.9, highlighting the need for careful tuning in practical deployments. Future research could explore GTPO's integration with larger models or other tool-based tasks, potentially unlocking new capabilities in AI reasoning. For now, this study sets a new standard in RL for TIR, emphasizing the importance of fine-grained rewards and temporal dynamics in training smarter, more adaptive AI systems.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn