Poster
in
Workshop: 1st Workshop on Foundation Models for Structured Data (FMSD)
State-Space Models for Tabular Prior-Data Fitted Networks
Felix Koch · Marcel Wever · Fabian Raisch · Benjamin Tischler
Recent advancements in foundation models for tabular data, such as TabPFN, demonstrated that pretrained Transformer architectures can approximate Bayesian inference with high predictive performance. However, Transformers suffer from quadratic complexity with respect to sequence length, motivating the exploration of more efficient sequence models. In this work, we investigate the potential of using Hydra, a bidirectional linear-time structured state space model (SSM), as an alternative to Transformers in TabPFN. A key challenge lies in SSM’s inherent sensitivity to the order of input tokens – an undesirable property for tabular datasets where the row order is semantically meaningless. We investigate to what extent a bidirectional approach can preserve efficiency and enable symmetric context aggregation.Our experiments show that this approach reduces the order-dependence, achieving predictive performance competitive to the original TabPFN model.