Poster
Communicating Activations Between Language Model Agents
Vignav Ramesh · Kenneth Li
East Exhibition Hall A-B #E-2604
Large language models can better reason through hard problems when multiple instances of the model (called “agents”) think through diverse approaches and communicate with each other, i.e. send each other messages in plain text. We wondered if LLM agents could communicate more efficiently and effectively by tapping into each other’s internal “thoughts” – a.k.a. the “activation vectors” produced as a model computationally processes a prompt. Our method pauses one model mid-computation, merges its activation with another model’s, and then continues processing. Crucially, this requires no extra training data or new model parameters. We evaluated our technique on multi-player coordination tasks and reasoning benchmarks, seeing a 27% boost in performance compared to text-based communication. Even better, this requires less than a fourth of the compute. These findings suggest that by sharing rich internal signals instead of words, LLM agents can collaborate far faster and more efficiently.