HuggingSara commited on
Commit
016cc32
·
1 Parent(s): 0f4b1a2

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - competitions/aiornot
4
+ language:
5
+ - en
6
+ metrics:
7
+ - accuracy
8
+ pipeline_tag: image-classification
9
+ ---
10
+
11
+ # Model Card for Model Soups on AirorNot Dataset
12
+
13
+ ## Model Details
14
+
15
+ ### Model Description
16
+
17
+ Code implementation of the paper [Model soups: averaging weights of multiple fine-tuned models
18
+ improves accuracy without increasing inference time](https://arxiv.org/abs/2203.05482).
19
+
20
+ In recent years, finetuning large models has been proving to be an excellent strategy to achieve high-performances in downstream tasks.
21
+ The conventional recipe to do so, it's to fine-tune models with different hyperparameters and select the one achieving the highest accuracy. However Wortsman *et. al* proved that averaging the weights of multiple models finetuned with different hyperparameter configurations can actually inprove accuracy and robustness.
22
+
23
+ I read this paper recently and I felt intrigued by the powerful yet simple idea (achieving a SOTA result on Imagenet of s 90.94%) so I decided that this could be an opportunity to get my hands dirty and dive into the code and...try the soup!
24
+
25
+ I started by using the official [code implementation](https://github.com/mlfoundations/model-soups) with CLIP ViT-B/32 and finetuned only 5 of their models on AiorNot. I used a simple strategy with minimal modifications. Mainly, I finetuned the models for 8 epochs with a batch size of 56 samples.
26
+
27
+ The tricky part was that I had to modify the baseline to use it with our custom dataset.
28
+
29
+ To implement this notebook I modified the version by [Cade Gordon](https://cadegordon.io/).
30
+
31
+
32
+ - **Developed by:** HuggingSara
33
+ - **Model type:** Computer Vision
34
+ - **Language :** Python
35
+
36
+
37
+