Eye On A.I. cover image

#123 Aidan Gomez: How AI Language Models Will Shape The Future

Eye On A.I.

CHAPTER

The Multi-Layer Perceptron for Language Learning

Multi-layer perceptron sounds like a neural, deep nut, but. That's the fundamental unit. And before transformers, there were these very complicated LSTM architecture with gates and all of these like confusing bits and bobs that just made it work. With the transformer, all of that was torn away. And the layer became MLPs plus one attention. It turns out to be extraordinarily powerful. The architecture is not this hyper complex beast. It's actually just a very simple, scalable compute saturating.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner