
Dynamic Token Merging for Efficient Byte-level Language Models with Julie Kallini - #724
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
00:00
Intro
This chapter explores the complex issues surrounding tokenization in language models, emphasizing the unequal impact on diverse languages. It also features the journey of a PhD student merging insights from computer science and linguistics in the field of natural language processing.
Play episode from 00:00
Transcript


