Spotlight Poster
Learning Parametric Distributions from Samples and Preferences
Marc Jourdan · Gizem Yüce · Nicolas Flammarion
West Exhibition Hall B2-B3 #W-804
Many AI systems today learn from examples, but newer methods are starting to use preferences---comparisons between options---to improve learning. We asked: When and how can preference feedback help models learn faster?In our study, we built a framework where a learner sees pairs of examples and is told which one is better. We found that this kind of feedback can significantly sharpen the model's estimates, especially when preferences are consistent and deterministic. In the best case, learning becomes much faster: the error drops much more quickly than when using examples alone.We also proved that this speed-up is the best possible under certain conditions. While our results rely on specific assumptions, they hold in important practical cases---like when AI models use scores based on how likely something is to happen.