AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
The Importance of Interpretability in Large Language Models
With convolutional neural networks, interpretability had many good advances. We can kind of understand the different layers have different sort of ranges that they're looking at. With large language models, it's perhaps a little more complicated. But I think it's still achievable in the sense that we could kind of ask, well, what kind of prompts this degenerate if I sort of drop out this part of the network? And sort of start getting at a language to even describe these types of aspects of human behavior or psychology from the spoken part in the language bar.