Personalized Benchmarking: Evaluating LLMs by Individual Preferences
Researchers propose a new method for evaluating LLMs that considers individual user preferences, moving beyond aggregate benchmarks. This approach uses ELO ratings to rank models based on personal context and needs.

A new paper from arXiv challenges the current approach to evaluating large language models (LLMs) by advocating for personalized benchmarks. The study highlights that existing benchmarks average preferences across all users, which can overlook the diverse needs and contexts of individual users. This one-size-fits-all approach may not accurately reflect how different users perceive model performance.
The researchers argue that personalized benchmarks are necessary because user preferences vary significantly depending on the context. For example, a technical writer might prioritize accuracy and coherence, while a creative writer might value originality and style. By using ELO ratings, a system commonly used in competitive games to rank players, the study demonstrates how models can be ranked based on individual user preferences, providing a more nuanced and relevant evaluation.
This shift towards personalized benchmarks could revolutionize how LLMs are developed and deployed. Companies might start tailoring models to specific user groups, enhancing satisfaction and effectiveness. However, implementing such a system at scale presents challenges, including the need for robust data collection and sophisticated algorithms to accurately capture individual preferences. The research opens up new avenues for understanding and improving LLM performance in real-world applications.