Poster
How Distributed Collaboration Influences the Diffusion Model Training? A Theoretical Perspective
Jing Qiao · Yu Liu · YUAN YUAN · Xiao Zhang · Zhipeng Cai · Dongxiao Yu
West Exhibition Hall B2-B3 #W-1002
Diffusion models are widely used to generate high-quality data, but most theories assume that the training process is based on only one powerful machine with abundant data.In practice, diffusion models often need to be trained across many devices that differ widely in speed, memory and local datasets, breaking the usual guarantee of accurate score estimation on each worker. We address this by proving the first generation-error bound for diffusion models trained in such resource-limited, heterogeneous settings. Remarkably, our bound shows a linear relationship with the data dimension—matching the best known single-machine results. We also show how key hyperparameters (such as learning rates, noise schedules and update frequencies) directly influence this bound and hence the final generation quality. By tuning these settings, we can balance workloads across devices and still achieve reliable, high-quality output.