Spotlight Poster
SAFE: Finding Sparse and Flat Minima to Improve Pruning
Dongyeop Lee · Kwanhee Lee · Jinseok Chung · Namhoon Lee
East Exhibition Hall A-B #E-3510
Pruning aims to reduce the memory and computational load of neural networks by zeroing out parameters, but often at the cost of accuracy. The core challenge lies in improving performance during pruning—or better yet, designing methods that can progress towards optimal performance. Such objectives are best studied in optimization, and recent advancements have shed light on new insights and strategies.Notably, recent research has emphasized the benefits of flat minima—regions in the loss landscape with low curvature. Intuitively, it can be understood as finding a solution where the changes in the loss or shifts in the parameters are more tolerable.This has not only shown to improve robustness and generalization in many prior studies, but it can potentially be highly desirable when the parameters are modified through pruning.This motivates us to propose SAFE, a constrained optimization algorithm that aims to find better sparse solutions through enforcing flatness, where we employ various established optimization techniques to enforce flatness while gradually imposing sparsity. We further extend this into SAFE+, which allows for flexible support of diverse pruning scores within its constrained optimization framework.Our results show that SAFE and SAFE+ successfully induce flatter and sparser minima, improving over existing baselines in both image classification and post-training pruning of language models, as well as improving robustness towards label noise, common image noises, and adversarial attacks.