Poster
FSL-SAGE: Accelerating Federated Split Learning via Smashed Activation Gradient Estimation
Srijith Nair · Michael Lin · Peizhong Ju · Amirreza Talebi · Elizabeth Bentley · Jia (Kevin) Liu
West Exhibition Hall B2-B3 #W-517
With the impressive performance of recently developed AI models like ChatGPT, there is a clear need for multiple computers or devices to take part in collaboratively training big models. Although many small devices like smartphones or smaller laptop computers have a lot of useful data to train large AI models, they are unable to participate in training because they have much lower resources like GPU, memory, etc. We design an algorithm that can allow such low-resource devices to take part in training large models without breaching the privacy of their data, with whatever resources they can provide for doing so. Our method and its future extensions can greatly shape the way we train AI models today, by enabling collaboration between many small-resource players with useful data, like academic institutions, hospitals, small firms, etc., while preserving privacy standards.