Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Methods and Opportunities at Small Scale (MOSS)

Generative or Discriminative? Revisiting Text Classification in the Era of Transformers

Siva Rajesh Kasa · Sumegh Roychowdhury · Karan Gupta · Yaswanth Biruduraju · SANTHOSH KASA · Ashutosh Kumar · Pattisapu Priyatam · Arindam Bhattacharya · Shailendra Agarwal · Vijay huddar

Keywords: [ Autoregressive models ] [ Encoder Models ] [ Generative Classifiers ] [ Discrete Diffusion Models ] [ Masked Language Models ]


Abstract:

In text classification, the classical comparison between discriminative and generative classifiers gains renewed relevance in the transformer era, where computational constraints often limit thorough experimentation. Through systematic small-scale experiments on text classification tasks, we investigate how the fundamental ``two regimes" phenomenon—where generative classifiers excel with limited data but show higher asymptotic error—manifests across modern architectures (Auto-regressive, Masked Language Models, Discrete Diffusion, and Encoders). By training models from scratch on controlled text datasets, we isolate and analyze core architectural behaviors in terms of sample efficiency, calibration, and preservation of ordinal relationships. Our findings provide insights into the inherent trade-offs of different modelling approaches for text classification, demonstrating how small-scale experimentation can inform both theoretical understanding and practical architectural choices.

Chat is not available.