Poster
Scaling Video-Language Models to 10K Frames via Hierarchical Differential Distillation
CHUANQI CHENG · Jian Guan · Wei Wu · Rui Yan
East Exhibition Hall A-B #E-3306
Long-form video processing fundamentally challenges vision-language models (VLMs) due to the high computational costs of handling extended temporal sequences. Existing token pruning and feature merging methods often sacrifice critical temporal dependencies or dilute semantic information. We introduce differential distillation, a principled approach that systematically preserves task-relevant information while suppressing redundancy. Based on this principle, we develop ViLAMP, a hierarchical video-language model that processes hour-long videos at "mixed precision" through two key mechanisms: (1) differential keyframe selection that maximizes query relevance while maintaining temporal distinctiveness at the frame level and (2) differential feature merging that preserves query-salient features in non-keyframes at the patch level. Hence, ViLAMP retains full information in keyframes while reducing non-keyframes to their most salient features, resembling mixed-precision training. Extensive experiments demonstrate ViLAMP's superior performance across five video understanding benchmarks, particularly on long-form content. Notably, ViLAMP can process ultra-long videos (up to 10K frames) on a single NVIDIA A100 GPU, achieving substantial computational efficiency while maintaining state-of-the-art performance. Code and model are available at https://github.com/steven-ccq/ViLAMP.
We developed ViLaMP, an AI model that can understand videos up to three hours long—all using just a single standard graphics card (A100 GPU). The idea is inspired by how humans watch videos: we pay close attention to important scenes and quickly skim through the rest.ViLaMP does something similar by using two smart techniques: (1) It identifies the most important moments in a video based on the task at hand and (2) It summarizes less important parts without losing their meaning. With these techniques, ViLaMP not only reduces computing costs but also beats other models on five major video understanding benchmarks. This makes it a practical and accurate tool for analyzing long videos, striking the right balance between detail and efficiency.