Poster
KIND: Knowledge Integration and Diversion for Training Decomposable Models
Yucheng Xie · Fu Feng · Ruixiao Shi · Jing Wang · Yong Rui · Xin Geng
East Exhibition Hall A-B #E-3501
Modern AI models often rely on pre-trained backbones, but these models are typically fixed in size and don’t adapt well when applied to different tasks—especially if there’s a big difference between training and deployment scenarios. To overcome this, we introduce KIND, a new pre-training approach that builds models out of smaller, reusable pieces. KIND uses a mathematical technique called Singular Value Decomposition (SVD) to break knowledge into basic components. These components are grouped into learngenes, which carry general-purpose knowledge, and tailors, which adapt to specific tasks. During training, a mechanism called a “class gate” separates these two types of knowledge. Experiments show that models trained with KIND can be flexibly adapted to different devices or domains by recombining the components. In particular, transferring just the general-purpose learngenes proves effective in new tasks with limited resources or large domain differences. Code is available at: https://github.com/Te4P0t/KIND.