Poster
in
Workshop: ES-FoMo III: 3rd Workshop on Efficient Systems for Foundation Models
Multi-student Diffusion Distillation for Better One-step Generators
Yanke Song · Jonathan Lorraine · Weili Nie · Karsten Kreis · James Lucas
Abstract:
Foundation models that generate photorealistic video with text or image guidance promise compelling augmented‑reality (AR) experiences, yet their prohibitive test‑time compute prevents true real‑time deployment. We focus on the dominant diffusion family and show that Multi-Student Distillation (MSD) increases effective model capacity without increasing — or even reducing — latency, memory footprint, or energy per sample. MSD partitions the conditioning space and trains a lightweight one‑step generator per partition, allowing (i) higher sample quality at fixed latency and (ii) smaller per‑student backbones that for edge/low-latency budgets.
Chat is not available.