AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
The Failure of Alignment in GPT 3
I think it would be much easier if both of you had like concrete Proposals for alignment and you just have like the pseudocode for both of you. You're asking like what happens as this thing that gets like dangerously smart And that is not going to be transparent in the code Let me come back to that on your first point about these things you know the alignment not generalizing Given that you've updated in the direction where the same sort of stacking more layers on the More attention layers is going to work. It seems that there will be more generalization between like GPD 4 and GPD 5 So I mean presumably whatever alignment techniques you used on GPD 2 would have worked on