Gradient Dissent: Conversations on AI cover image

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

Gradient Dissent: Conversations on AI

CHAPTER

The Importance of Attention in Recurrent Neural Networks

I remember when attention first came out on these recurrent neural networks. I guess what inspired you to try the attention mechanism without the underlying LSTM? So it's a very good thing to succeed.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner