AIResearch AIResearch
Back to articles
AI

AI Judges User Preferences Without Training

Large language models can now evaluate complex recommendation systems by comparing item lists, offering a lightweight alternative to traditional simulators while maintaining logical consistency.

AI Research
November 08, 2025
3 min read
AI Judges User Preferences Without Training

The algorithms that determine what we watch, read, and buy are becoming increasingly sophisticated, yet evaluating their performance remains a fundamental challenge. A new study demonstrates how large language models (LLMs) can act as impartial judges of user preferences, providing a practical method for assessing recommendation systems without requiring extensive user data or complex simulations.

Researchers discovered that pretrained LLMs can effectively articulate preferences between different recommendation slates—ordered sequences of items like movies, products, or songs. By framing evaluation as pairwise comparisons between candidate slates, the models demonstrated the ability to capture meaningful user preference structures. The key finding reveals that LLMs maintain strong logical consistency in their judgments, with transitivity rates reaching up to 0.997 across different tasks, indicating that when the model prefers slate A over B and B over C, it consistently prefers A over C.

The methodology employed a novel LLM-as-a-Judge approach where models compared pairs of recommendation slates without any task-specific training. Each evaluation followed a structured four-part prompt: introducing the recommendation context, providing user interaction history, presenting two candidate slates with their item features, and requiring a forced-choice selection between them. To mitigate positional bias, each slate pair was evaluated twice with reversed order, and results were aggregated through majority voting across multiple model queries.

Analysis of results across three recommendation tasks and five datasets revealed distinct performance patterns. For slate selection tasks (choosing which items to recommend), LLMs achieved empirical regret values as low as 0.044, significantly outperforming random baselines. In ordering tasks (determining how to arrange items), performance was more challenging with regret values around 0.2, reflecting the difficulty of fine-grained preference judgments between similar slates. Most notably, in joint selection and ordering tasks—the most realistic scenario—LLMs demonstrated strong performance with regret values as low as 0.005, suggesting they excel when both content selection and arrangement matter.

This approach matters because it offers a practical alternative to traditional recommendation system evaluation, which typically requires either live user testing—expensive and slow—or complex simulators that model fine-grained user behavior. The LLM-based method provides immediate, scalable evaluation that can transfer across domains from e-commerce to entertainment, enabling faster iteration and improvement of recommendation algorithms that shape our digital experiences.

The study acknowledges limitations in the models' ability to handle highly similar slates that differ only in item ordering, where performance gaps between LLMs and baselines narrow. Additionally, while the models demonstrate strong logical consistency, their preference articulations may not perfectly align with all aspects of human judgment, particularly for nuanced cultural or contextual factors not captured in the training data.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn