
How Does ChatGPT Work? - ML 107
Adventures in Machine Learning
00:00
How to Implement Chat GPT in a Hugging Phase
The architecture is a decoder-only transformer network with a 2048 token long context. Within that, it has 175 billion parameters, which is just completely unprecedented in this space. This architecture to put it into context like if you wanted to get the performance out of chat that we currently see in chat GPT three or in the far more advanced DaVinci.
Transcript
Play full episode