ViPO: Visual Preference Optimization at Scale

📰 ArXiv cs.AI

arXiv:2604.24953v2 Announce Type: cross Abstract: While preference optimization is crucial for improving visual generative models, how to effectively scale this paradigm remains largely unexplored. Current open-source preference datasets contain conflicting preference patterns, where winners excel in some dimensions but underperform in others. Naively optimizing on such noisy datasets fails to learn preferences, hindering effective scaling. To enhance robustness against noise, we propose Poly-DP

Published 29 Apr 2026
Read full paper → ← Back to Reads