Model Card for Model Soups on AirorNot Dataset
Model Details
Model Description
Code implementation of the paper Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time.
In recent years, finetuning large models has been proving to be an excellent strategy to achieve high-performances in downstream tasks. The conventional recipe to do so, it's to fine-tune models with different hyperparameters and select the one achieving the highest accuracy. However Wortsman et. al proved that averaging the weights of multiple models finetuned with different hyperparameter configurations can actually inprove accuracy and robustness.
I read this paper recently and I felt intrigued by the powerful yet simple idea (achieving a SOTA result on Imagenet of s 90.94%) so I decided that this could be an opportunity to get my hands dirty and dive into the code and...try the soup!
I started by using the official code implementation with CLIP ViT-B/32 and finetuned only 5 of their models on AiorNot. I used a simple strategy with minimal modifications. Mainly, I finetuned the models for 8 epochs with a batch size of 56 samples.
The tricky part was that I had to modify the baseline to use it with our custom dataset.
- Developed by: HuggingSara
- Model type: Computer Vision
- Language : Python