Spotlight Poster
Upweighting Easy Samples in Fine-Tuning Mitigates Forgetting
Sunny Sanyal · Hayden Prairie · Rudrajit Das · Ali Kavis · Sujay Sanghavi
East Exhibition Hall A-B #E-1302
Pretrained AI models are powerful, but fine-tuning them on new tasks often causes them to forget what they originally learned — a problem known as catastrophic forgetting. This is especially challenging when the original training data is unavailable, as is often the case with publicly released models.To address this, we propose a simple method called FLOW. Unlike traditional approaches focusing on difficult examples during fine-tuning, FLOW does the opposite: it upweights “easy” examples — cases that the model already handles well. This helps preserve the model’s original capabilities while still allowing it to adapt to the new task.FLOW is easy to use, requires no extra parameters, and relies only on a basic calculation using the model’s pre-trained performance. Despite its simplicity, it consistently outperforms standard fine-tuning and other relevant baselines across both vision and language tasks. For example, when fine-tuning for math reasoning, FLOW achieved similar accuracy on new math problems while retaining 5% more accuracy on the original general knowledge tasks than standard fine-tuning.Because FLOW doesn’t require access to original training data, it offers a practical solution for retaining valuable model capabilities while fine-tuning on new tasks.