Adventures in Machine Learning cover image

How Does ChatGPT Work? - ML 107

Adventures in Machine Learning

00:00

How to Implement Chat GPT in a Hugging Phase

The architecture is a decoder-only transformer network with a 2048 token long context. Within that, it has 175 billion parameters, which is just completely unprecedented in this space. This architecture to put it into context like if you wanted to get the performance out of chat that we currently see in chat GPT three or in the far more advanced DaVinci.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app