Machine Learning Street Talk (MLST) cover image

Transformers Need Glasses! - Federico Barbero

Machine Learning Street Talk (MLST)

00:00

Exploring Machine Learning Models and Their Challenges

This chapter investigates two key experiments on machine learning models, focusing on the impact of sequence interleaving and representation on performance. It highlights the struggles of models with long sequences, particularly in maintaining accurate attention during arithmetic tasks, and the reliance on heuristics for counting. The conversation also covers the theoretical limitations of transformers, their integration with specialized modules, and the trade-offs involved in navigating between connectionist and hybrid approaches in AI.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app