Machine Learning Street Talk (MLST) cover image

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

Machine Learning Street Talk (MLST)

00:00

Transformers and Positional Encoding

This chapter explores the intricacies of feature transformation, metric learning, and the innovative use of positional encoding in transformer architectures. It contrasts traditional recurrent neural networks with transformers, focusing on their ability to handle long-term dependencies and the significance of relative embeddings. The discussion also touches on the T5 model's learnable scaler encoding and the impact of unsupervised objectives on self-supervised language models.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app