
121. Alexei Baevski - data2vec and the future of multimodal learning
Towards Data Science
Using Latent Representations as a Target for Your Model Training
The traditional task that i've talked about before is a reconstructing the original into data. This works reasonably well, but then you have to makea choice. What we found actually, well for wivh to bacc, is trying to using the latent representations as the target for your model to to preject. So we use this with contrestive learning. And algertings like b el and ina show that you can use a similar approach by reconstructing latent representations as a way to train your model. The problem with at is, if you're training the model t predict what the model is producing, then the model will very, very quickly a collapse into producing the
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.