Large language models like ChatGPT and Claude learn what responses humans prefer through a process called reinforcement learning from human feedback (RLHF). At the heart of this system lies a reward model—a specialized AI that scores how good a response is, guiding the main model toward better answers. However, these reward models have struggled with two persistent problems: they often waste training data by not clearly understanding their scoring task, and they can be easily tricked into giving high scores to poor responses, a flaw known as reward overoptimization. A new approach called PIRA tackles these issues head-on, offering a more stable and efficient way to train the judges that teach AI how to behave.
The researchers discovered that by explicitly telling the reward model what to look for in each evaluation, they could dramatically improve its accuracy and consistency. Traditional s simply feed the question and answer into the model without clarification, forcing it to guess that its job is to score preferences. PIRA reformats this input by adding clear instructions like "Evaluate whether the response demonstrates a reliable grasp of facts and reasoning" or "Judge the extent to which the response stays aligned with the user's intent." This simple change makes the task explicit, allowing the model to learn more effectively from the same amount of data. The paper shows that this preference-oriented instruction reformulation alone boosted accuracy on the HH-cleaned dataset from 64.3% to 73.0% when using the LLaMA3-8B model.
To build PIRA, the team developed a three-part ology. First, they created a set of ten diverse evaluation instructions through a combination of AI generation and human review. These instructions cover different aspects of response quality, from factual accuracy and logical coherence to clarity and utility. During training, for each question-answer pair, the model randomly receives one of these instructions prepended to the input, teaching it to evaluate responses from multiple perspectives. Second, during inference, PIRA doesn't rely on a single instruction but averages scores across multiple instructions—typically six different ones—to reduce bias from any particular phrasing. Third, it applies a technique called stochastic value-head averaging, where the final scoring layer runs multiple times with slight random variations (dropout) and averages those , creating a more stable reward estimate without significantly increasing computation time.
Demonstrate clear improvements across multiple benchmarks. As shown in Table 1 of the paper, PIRA consistently outperformed baseline s and previous approaches like Thomas across six different datasets (HH, Oasst, SHP, UltraFeedback, Alpaca-farm, and HH-cleaned) and four model architectures (Mistral-7B, LLaMA3-8B, and two sizes of Qwen2.5). On average, PIRA achieved 70.1% accuracy with Mistral-7B compared to 65.7% for the baseline, and similar gains appeared with other models. Perhaps more importantly, PIRA effectively mitigated reward hacking—the phenomenon where AI systems learn to exploit flaws in the reward model to get high scores while producing worse responses. Figures 1 and 2 show that while baseline s exhibited sharp spikes in reward inflation followed by collapse, PIRA maintained stable, monotonic improvement in gold reward scores throughout training.
Of this research extend to making AI systems more reliable and trustworthy. By improving how reward models learn, PIRA helps ensure that language models align more closely with genuine human preferences rather than finding shortcuts to high scores. This matters for everyday users because it means AI assistants are less likely to produce misleading, inconsistent, or unhelpful responses. also shows particular strength in low-data scenarios—Figure 4 reveals that with only 500 training examples, PIRA improved accuracy by approximately 9 percentage points over the baseline, suggesting it could make AI training more data-efficient. Additionally, the cross-dataset evaluation in Figure 5 demonstrates that PIRA generalizes well to new types of questions and responses, maintaining robustness even when the training and testing data come from different sources.
Despite these advances, PIRA has limitations that the researchers acknowledge. hasn't yet been tested on language models larger than 8 billion parameters, so its scalability to today's frontier models remains uncertain. The dual aggregation approach—averaging across multiple instructions and stochastic forward passes—introduces some computational overhead, though the paper notes this is minimal for the value-head averaging (about 7% increased latency with 12 passes) but more substantial for instruction-set averaging. Furthermore, the effectiveness decreases with very long or complex responses, as seen in the UltraFeedback dataset , indicating that response length can still the approach. These limitations point to areas for future refinement as AI systems continue to evolve.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn