A research team from Google proposes ByT5 architecture, a competitive token-free pretrained byte-to-byte transformer that can be straightforwardly adapted to process byte sequences without adding excessive computational cost.

Here is a quick read: Towards a Token-Free Future: Google Proposes Pretrained Byte-to-Byte Transformers for NLP.

The paper ByT5: Towards a Token-Free Future With Pre-Trained Byte-to-Byte Models is on arXiv.

Source link