Gradient Dissent: Conversations on AI cover image

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

Gradient Dissent: Conversations on AI

00:00

The Importance of Attention in Recurrent Neural Networks

I remember when attention first came out on these recurrent neural networks. I guess what inspired you to try the attention mechanism without the underlying LSTM? So it's a very good thing to succeed.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app