QUIVER: Cost-Aware Adaptive Preference Querying in Surrogate-Assisted Evolutionary Multi-Objective Optimization
TLDR
QUIVER is a cost-aware, adaptive multi-objective optimizer that intelligently balances objective evaluations and heterogeneous preference queries to minimize regret.
Key contributions
- Introduces QUIVER, an adaptive surrogate-assisted evolutionary multi-objective optimizer.
- Maximizes expected decision-quality improvement per unit total cost for action selection.
- Adaptively chooses between objective evaluations and heterogeneous preference queries (PS/IA).
- Achieves 25% lower utility regret on challenging WFG problems, outperforming baselines.
Why it matters
This paper addresses the critical budget allocation dilemma in interactive multi-objective optimization. By adaptively balancing expensive evaluations and diverse preference queries, QUIVER significantly improves decision quality and efficiency. It offers a novel approach to cost-aware preference learning.
Original Abstract
Interactive multi-objective optimization systems face a budget allocation dilemma: one can spend resources on expensive objective evaluations or on eliciting decision-maker preferences that identify the relevant region of the Pareto set. Moreover, preference elicitation itself spans modalities with different information content and cognitive burden, ranging from cheap, noisy pairwise preference statements (PS) to richer but costlier indifference adjustments (IA). We study cost-aware optimization under an unknown scalarization and introduce QUIVER (Query-Informed Value Estimation for Regret), a surrogate-assisted evolutionary multi-objective optimizer that adaptively chooses between objective evaluations and heterogeneous preference queries. At each step, QUIVER selects the next action by maximizing the expected decision-quality improvement per unit total cost. Across DTLZ and WFG benchmarks under synthetic decision-maker models, QUIVER achieves the lowest final utility regret on challenging WFG problems (utility regret of 2.14 on WFG4, 2.82 on WFG9: a 25% improvement over baselines), outperforming all single-modality baselines. We analyze how the optimal mix of PS and IA adapts to problem difficulty: on easy problems (DTLZ2), QUIVER selects 80\% PS queries; on hard problems (WFG9), it shifts to 35% IA queries. This adaptive modality selection demonstrates cost-aware preference learning in action.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.