|
--- |
|
datasets: |
|
- competitions/aiornot |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
pipeline_tag: image-classification |
|
--- |
|
|
|
# Model Card for Model Soups on AirorNot Dataset |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
Code implementation of the paper [Model soups: averaging weights of multiple fine-tuned models |
|
improves accuracy without increasing inference time](https://arxiv.org/abs/2203.05482). |
|
|
|
In recent years, finetuning large models has been proving to be an excellent strategy to achieve high-performances in downstream tasks. |
|
The conventional recipe to do so, it's to fine-tune models with different hyperparameters and select the one achieving the highest accuracy. However Wortsman *et. al* proved that averaging the weights of multiple models finetuned with different hyperparameter configurations can actually inprove accuracy and robustness. |
|
|
|
I read this paper recently and I felt intrigued by the powerful yet simple idea (achieving a SOTA result on Imagenet of s 90.94%) so I decided that this could be an opportunity to get my hands dirty and dive into the code and...try the soup! |
|
|
|
I started by using the official [code implementation](https://github.com/mlfoundations/model-soups) with CLIP ViT-B/32 and finetuned only 5 of their models on AiorNot. I used a simple strategy with minimal modifications. Mainly, I finetuned the models for 8 epochs with a batch size of 56 samples. |
|
|
|
The tricky part was that I had to modify the baseline to use it with our custom dataset. |
|
|
|
|
|
- **Developed by:** HuggingSara |
|
- **Model type:** Computer Vision |
|
- **Language :** Python |
|
|
|
|
|
|
|
|