Poster
Exactly Tight Information-theoretic Generalization Bounds via Binary Jensen-Shannon Divergence
Yuxin Dong · Haoran Guo · Tieliang Gong · Wen Wen · Chen Li
West Exhibition Hall B2-B3 #W-919
Machine learning models often perform well on the data they’re trained on, but the real challenge is ensuring they do just as well on new, unseen data. To understand and improve this “generalization” ability, researchers have developed mathematical tools called generalization bounds. These bounds try to measure how far off a model’s performance on training data might be from its performance on future data.However, existing tools sometimes give very loose estimates. This paper introduces a new way to get much sharper, more accurate estimates. We focus on a well-known information-theoretic measure, but simplify it using a “binary” version, breaking complex outcomes down to simpler yes/no signals. This makes the math more manageable and leads to better bounds.Moreover, through a new technique called “binarization”, we provide the first generalization bounds that are not only more accurate but exactly tight: matching the best possible performance for a wide range of machine learning methods, without needing extra assumptions. This makes our bounds particularly valuable for analyzing modern, randomized learning algorithms used in areas like deep learning and optimization.In short, this work improves our ability to trust machine learning models by offering stronger, more precise tools to measure how well they will generalize to new data.