

#43030
Mentioned in 1 episodes
Barlow Twins: Self-Supervised Learning via Redundancy Reduction
A Novel Approach to Self-Supervised Learning
Book • 2021
Barlow Twins is a self-supervised learning approach that leverages redundancy reduction to create embeddings invariant to input distortions.
It uses two identical neural networks to minimize redundancy between vector components, ensuring robust representations.
This method is competitive with state-of-the-art self-supervised learning techniques.
It uses two identical neural networks to minimize redundancy between vector components, ensuring robust representations.
This method is competitive with state-of-the-art self-supervised learning techniques.
Mentioned by
Mentioned in 1 episodes
Mentioned by 

when discussing self-supervised contrastive learning papers.


Tim Scarfe

21 snips
#55 Self-Supervised Vision Models (Dr. Ishan Misra - FAIR).