Poster
Generalization and Robustness of the Tilted Empirical Risk
Gholamali Aminian · Amir R. Asadi · Tian Li · Ahmad Beirami · Gesine Reinert · Samuel Cohen
West Exhibition Hall B2-B3 #W-907
When we train a machine-learning model we want to know how well it will perform on new, unseen data—this is called its generalization error. Researchers estimate that error by averaging how wrong the model is on the training set.Recently, a technique called the tilted empirical risk (TER) was proposed. Instead of giving every training example equal weight, TER lets you tilt the calculation so that unusually large errors are emphasized less (negative tilt). In this paper, we answer the following question: What happens if we use negative tilt with “outliers” ?We show two main results:1- Reliable performance guarantees without data shift: We prove mathematical limits on how far TER can be different from the unknown error even when individual errors (losses) can become very large. As the training set grows, that gap shrinks at a predictable pace, meaning TER remains a reliable method.2- Robustness when the data are noisy: We analyze scenarios where the training data are contaminated with noisy outliers or come from a slightly different distribution than the data the model will face later (a “distribution shift”). The theory shows that, under negative tilt, TER still gives dependable guidance.Finally, we run simple experiments to confirm the theory and to demonstrate a practical way to pick the best amount of tilt directly from the data.In short, the study explains why—and by how much—down-weighting extreme training errors (using negative tilt) can make risk estimates both stable and resilient to bad (noisy) data, giving practitioners a solid, data-driven method to tune when their datasets are noisy.