Poster
Sorbet: A Neuromorphic Hardware-Compatible Transformer-Based Spiking Language Model
Kaiwen Tang · Zhanglu Yan · Weng-Fai Wong
East Exhibition Hall A-B #E-3203
To protect user privacy, we aim to run language models directly on small devices like phones, which have limited computing power and need to save energy. Many key steps in standard language models are hard to run efficiently on such low-power hardware. To solve this, we developed Sorbet, a new language model based on spiking neural networks (SNN) that is designed specifically for energy-efficient hardware. Sorbet replaces traditional power-hungry operations with new methods, making the model much more energy-saving. Using advanced techniques, we also compressed Sorbet into a very small model while keeping strong accuracy. Our tests show that Sorbet uses over 27 times less energy than typical models. This work opens the door to smarter, more energy-efficient devices that better protect user privacy.