Publications

Language models (LMs) are bound to their tokenizer, which maps raw text to a sequence of vocabulary items (tokens). This restricts …

Transformers have emerged as the backbone of large language models (LLMs). However, generation remains inefficient due to the need to …