ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models
Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel
March 2022Abstract
ByT5 studies byte-level sequence modeling at scale and demonstrates strong multilingual performance without tokenization.
Publication
Transactions of the Association for Computational Linguistics

Member of Technical Staff - TLM
My research interests include language modeling, embodied AI, motion forecasting, and multilingual modeling.