
Dynamic Token Merging for Efficient Byte-level Language Models with Julie Kallini - #724
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Intro
This chapter explores the complex issues surrounding tokenization in language models, emphasizing the unequal impact on diverse languages. It also features the journey of a PhD student merging insights from computer science and linguistics in the field of natural language processing.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.