AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
The Importance of Interpretability in Neural Architectures
The paper uses those around the term feature, but everything's a feature in these systems to sub-dupery or another. Are they individual dimensions or are they pairs or tuples of dimensions that kind of carry the meaning here? That's a great question. So we started looking at individual dimensions and asking exactly as you say, like for pick a single element in the given embedding,. What can we learn about what that embedding captures among all the dimensions? And we can talk about how we go about doing that.