Poster
BSemiFL: Semi-supervised Federated Learning via a Bayesian Approach
Haozhao Wang · Shengyu Wang · Jiaming Li · Hao Ren · Xingshuo Han · Wenchao Xu · Shangwei Guo · Tianwei Zhang · Ruixuan Li
West Exhibition Hall B2-B3 #W-520
Semi-supervised federated learning is a method that allows multiple users (we call them "clients") to collaborate in training an AI model. The advantage of this approach is that even if the data on each client's device is unlabeled, and without sharing the actual data, all clients can still jointly train a reasonably effective global model.A key step in this process is called "re-labeling". In this step, each client uses two pre-trained models —— a global model shared by all users and a local model trained specifically on their own data —— to automatically assign labels to their unlabeled data. These newly labeled data are then used to further improve the model.Past research has tried various ways to perform this labeling process —— for example, using only the global model, or taking an average of predictions from both the global and local models. However, few studies have deeply explored the strengths and limitations of these two models.In our work, through theoretical analysis and experimental validation, we observed an important phenomenon:The local model is better suited for labeling its own local data and achieves higher accuracy.The global model, although initially less accurate, gradually improves its labeling performance as it incorporates knowledge from other clients' data.Based on this finding, we propose a new method called BSemiFL. Its core idea is that instead of simply choosing one model or averaging their outputs, we use a technique called Bayesian inference to assess which model — the local or the global — is more "similar" to the current data sample. Based on this similarity, we compute weights for the two models and combine their predictions to generate more accurate pseudo-labels.Theoretically, we show that this method results in fewer labeling errors compared to using only the local model, only the global model, or simply averaging the two. Experimental results also demonstrate that, compared to the current state-of-the-art methods, our approach improves performance by up to 9.8%.