Papers Read on AI

Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time

Jun 28, 2024
The podcast discusses how averaging weights of multiple fine-tuned models can improve accuracy without increasing inference time. They introduce the concept of 'model soups,' which outperforms conventional ensemble methods. The approach is showcased on pre-trained models like CLIP, ALIGN, and ViT-G, leading to state-of-the-art results. Additionally, they explore the benefits of model soups in various tasks and analyze the relation between weight-averaging and flatness of loss.
Ask episode
Chapters
Transcript
Episode notes