Skip to yearly menu bar Skip to main content


Invited Talk
in
Workshop: Tokenization Workshop (TokShop)

Insights from Pixel Language Modeling

Desmond Elliott

[ ]
Fri 18 Jul 10 a.m. PDT — 10:50 a.m. PDT

Abstract:

Language models typically have a multilingual representation bottleneck starting from their tokenizers. As an alternative vision, I will talk about pixel language models, which avoid this bottleneck by rendering text as images. Pixel language models can transfer within and across writing systems without paying a computational cost for their flexibility. I will present a series of models, pretrained on only English text or with multilingual pretraining, at scales ranging from just 5M parameters up to 86M parameters, on downstream tasks in 14 scripts. I will conclude by reflecting on the lessons learned from the early years of this type of tokenization-free representation.

Chat is not available.