Gradient Dissent: Conversations on AI cover image

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

Gradient Dissent: Conversations on AI

00:00

The Future of Large Language Models

I do believe that there are a lot of possible architectures that would be fast efficient and result in performance that we're seeing from the current large language models. There are some things you can't literally just scale up a MLP was relose because that would be done point wise, right? You wouldn't be able to learn relationships between words. But so long as you're not breaking that or like severely compromising that, I think there's a huge swath of models that would perform equivalently well and would scale equivalently well.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app