Poster
On-Device Collaborative Language Modeling via a Mixture of Generalists and Specialists
Dongyang Fan · Bettina Messmer · Nikita Doikov · Martin Jaggi
East Exhibition Hall A-B #E-3309
Modern language models are increasingly deployed on personal devices to preserve user privacy and improve personalization. However, this approach faces two major challenges: devices differ in computational power (model heterogeneity), and users have unique data and language habits (data heterogeneity). Traditional methods cannot effectively address both at once.We introduce CoMiGS, a collaborative learning framework that blends shared "generalist" knowledge with user-specific "specialist" insights. It dynamically routes each word prediction to the most suitable expert using a novel bi-level optimization algorithm that separates training and validation phases.CoMiGS enables efficient, privacy-preserving language model customization on devices with varying capabilities. It reduces communication costs by 50%, minimizes risk of overfitting, and delivers consistent performance across users. This makes it a practical foundation for smarter, more adaptive AI on mobile and edge devices.