Poster
SyncMind: Measuring Agent Out-of-Sync Recovery in Collaborative Software Engineering
Xuehang Guo · Xingyao Wang · Yangyi Chen · Sha Li · Chi Han · Manling Li · Heng Ji
East Exhibition Hall A-B #E-1807
What happens when your teammates contribute new updates to your collaborative project while you're away, leaving you out of the loop? In collaborative scenarios, team members can fall "out-of-sync" with the current state of the project—missing crucial changes and potentially introducing bugs or causing integration issues. This problem becomes increasingly significant as more collaborative environments incorporate AI assistants that need to stay aligned with human collaborators.Our research tackles this challenge through the lens of collaborative software engineering, where we introduce SyncMind, a framework that systematically measures how well large language model (LLM) agents recover when they lose synchronization with the coding environment. We constructed SyncBench, a scalable benchmark with an open-source construction method to evaluate agents' ability in detecting, localizing, and fixing synchronization issues. Our experiments not only revealed striking performance and ability differences across LLM agents, but also provided insights into resource-efficient collaboration in future collaborative systems.As agents can fall out-of-sync in various collaborative scenarios beyond software engineering, our findings carry implications for broad collaborative systems—whether among humans, AIs, or mixed teams—highlighting the need for better collaborative mechanisms, improved resource awareness in AI assistants, as well as strategic reasoning, planning, and decision-making capabilities to help agents efficiently and effectively recover from state misalignments.