
107 - Multi-Modal Transformers, with Hao Tan and Mohit Bansal
NLP Highlights
00:00
The Multi-Modal Representation Is a Good Example
Ando: I think a carrently think its using the language repetitiontiov long tal probem. Because actually, if o we just take the un mode of petrini, the repetition of the word embodying the vision petin, and the language petriin,. they are very different to each other in detail. Ando: In the long term, when we figure this out, in some sense, that the multi modal representation should help us even on the language side, for language only tasks? That's like the long term goal of this whole area, right? Yes, i think so.
Transcript
Play full episode