File size: 30,017 Bytes
b09b05b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cf36132
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b09b05b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
---
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- feature-extraction
- sentence-similarity
- transformers

---

# {MODEL_NAME}

This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.

<!--- Describe your model here -->

## Usage (Sentence-Transformers)

Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:

```
pip install -U sentence-transformers
```

Then you can use the model like this:

```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]

model = SentenceTransformer('{MODEL_NAME}')
embeddings = model.encode(sentences)
print(embeddings)
```



## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.

```python
from transformers import AutoTokenizer, AutoModel
import torch


#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
    token_embeddings = model_output[0] #First element of model_output contains all token embeddings
    input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
    return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)


# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']

# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
model = AutoModel.from_pretrained('{MODEL_NAME}')

# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')

# Compute token embeddings
with torch.no_grad():
    model_output = model(**encoded_input)

# Perform pooling. In this case, mean pooling.
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])

print("Sentence embeddings:")
print(sentence_embeddings)
```



## Evaluation Results

| Model                                                  | id_raw_acc   | vn_raw_acc   | br_raw_acc   | th_raw_acc   | my_raw_acc   | ph_raw_acc   | sg_raw_acc   | avg        |
|:------------------------------------------------------------|:-------------|:-------------|:-------------|:-------------|:-------------|:-------------|:-------------|:-----------|
| [thtang/SetFit_ALL_200M_itr5](https://huggingface.co/thtang/SetFit_ALL_200M_itr5)                                 | **74.24%**       | **64.04%**   | **58.98%**       | **67.24%**       | **70.77%**       | **70.63%**   | **70.58%**       | **68.07%**     |
| [('ViT-B-16-SigLIP-i18n-256', 'webli')](https://huggingface.co/timm/ViT-B-16-SigLIP-i18n-256)                       | **69.38%**   | **57.92%**   | 47.40%       | **56.40%**   | 65.20%       | 65.72%       | 65.12%       | **61.02%** |
| ('xlm-roberta-base-ViT-B-32', 'laion5b_s13b_b90k')          | 66.23%       | 54.05%       | **49.26%**   | 55.39%       | **65.61%**   | 66.11%       | 66.72%       | 60.48%     |
| ('xlm-roberta-large-ViT-H-14', 'frozen_laion5b_s13b_b90k')  | 66.05%       | 52.77%       | 46.46%       | 53.44%       | 62.70%       | 64.40%       | 64.24%       | 58.58%     |
| ('ViT-L-14', 'commonpool_xl_s13b_b90k')                     | 65.48%       | 53.80%       | 46.61%       | 51.00%       | 62.01%       | 64.37%       | 63.94%       | 58.17%     |
| ('ViT-L-14', 'commonpool_xl_clip_s13b_b90k')                | 66.73%       | 49.82%       | 45.25%       | 38.32%       | 63.64%       | 66.17%       | 65.29%       | 56.46%     |
| ('ViT-B-16', 'commonpool_l_s1b_b8k')                        | 62.14%       | 49.25%       | 45.20%       | 39.47%       | 61.15%       | 63.03%       | 62.63%       | 54.69%     |
| ('ViT-bigG-14-CLIPA', 'datacomp1b')                         | 69.21%       | 44.39%       | 48.25%       | 20.54%       | 62.83%       | 68.15%       | 66.48%       | 54.26%     |
| ('ViT-bigG-14-CLIPA-336', 'datacomp1b')                     | 69.17%       | 44.22%       | 48.06%       | 20.48%       | 62.79%       | **67.74%**   | 66.63%       | 54.15%     |
| ('ViT-H-14-CLIPA-336', 'datacomp1b')                        | 68.03%       | 42.79%       | 47.52%       | 20.82%       | 62.38%       | 67.06%       | **66.92%**   | 53.65%     |
| ('ViT-H-14-CLIPA', 'datacomp1b')                            | 68.18%       | 42.82%       | 47.33%       | 20.68%       | 62.31%       | 67.26%       | 66.56%       | 53.59%     |
| ('ViT-B-16', 'commonpool_l_clip_s1b_b8k')                   | 63.68%       | 42.24%       | 44.87%       | 28.59%       | 62.04%       | 65.18%       | 64.97%       | 53.08%     |
| ('ViT-B-32-256', 'datacomp_s34b_b86k')                      | 65.44%       | 38.94%       | 43.57%       | 25.11%       | 62.39%       | 65.82%       | 64.94%       | 52.32%     |
| ('ViT-L-14-CLIPA-336', 'datacomp1b')                        | 66.99%       | 38.69%       | 45.25%       | 20.36%       | 61.47%       | 66.78%       | 65.56%       | 52.16%     |
| ('ViT-L-14-CLIPA', 'datacomp1b')                            | 66.86%       | 38.34%       | 45.21%       | 20.18%       | 61.51%       | 66.71%       | 65.41%       | 52.03%     |
| ('ViT-H-14-CLIPA-336', 'laion2b')                           | 64.62%       | 35.52%       | 44.73%       | 21.27%       | 61.01%       | 67.12%       | 65.76%       | 51.43%     |
| ('ViT-B-32', 'datacomp_xl_s13b_b90k')                       | 64.57%       | 37.26%       | 42.06%       | 22.61%       | 61.96%       | 65.59%       | 64.63%       | 51.24%     |
| ('ViT-L-14', 'datacomp_xl_s13b_b90k')                       | 64.37%       | 37.78%       | 40.65%       | 22.89%       | 60.72%       | 65.26%       | 64.30%       | 50.85%     |
| ('EVA02-E-14-plus', 'laion2b_s9b_b144k')                    | 63.51%       | 31.79%       | 42.52%       | 23.71%       | 60.74%       | 64.74%       | 63.97%       | 50.14%     |
| ('ViT-H-14-quickgelu', 'metaclip_fullcc')                   | 59.75%       | 34.61%       | 43.12%       | 22.69%       | 60.61%       | 65.47%       | 64.58%       | 50.12%     |
| ('ViT-B-16', 'datacomp_xl_s13b_b90k')                       | 63.15%       | 36.19%       | 39.81%       | 22.39%       | 60.66%       | 63.96%       | 63.31%       | 49.92%     |
| ('ViT-bigG-14', 'laion2b_s39b_b160k')                       | 63.03%       | 31.52%       | 41.20%       | 23.65%       | 60.52%       | 65.11%       | 63.99%       | 49.86%     |
| ('ViT-B-16', 'commonpool_l_basic_s1b_b8k')                  | 62.56%       | 36.99%       | 40.87%       | 22.16%       | 59.57%       | 63.56%       | 63.06%       | 49.82%     |
| intfloat/multilingual-e5-large                              | 52.99%       | 42.00%       | 33.92%       | 47.69%       | 55.82%       | 57.76%       | 58.16%       | 49.76%     |
| intfloat/multilingual-e5-base                               | 52.06%       | 43.21%       | 34.17%       | 47.41%       | 55.28%       | 57.38%       | 57.45%       | 49.57%     |
| ('ViT-B-16', 'commonpool_l_image_s1b_b8k')                  | 61.48%       | 36.08%       | 40.87%       | 22.62%       | 59.17%       | 63.47%       | 62.80%       | 49.50%     |
| ('convnext_large_d', 'laion2b_s26b_b102k_augreg')           | 61.61%       | 29.78%       | 39.92%       | 23.49%       | 60.93%       | 65.69%       | 64.60%       | 49.43%     |
| ('EVA01-g-14-plus', 'merged2b_s11b_b114k')                  | 62.34%       | 30.29%       | 39.02%       | 22.80%       | 60.83%       | 65.19%       | 63.49%       | 49.14%     |
| ('convnext_large_d_320', 'laion2b_s29b_b131k_ft')           | 61.18%       | 29.24%       | 39.09%       | 23.23%       | 60.65%       | 65.64%       | 64.12%       | 49.02%     |
| ('ViT-B-32', 'laion2b_s34b_b79k')                           | 61.21%       | 29.82%       | 37.51%       | 24.49%       | 60.21%       | 65.28%       | 64.08%       | 48.94%     |
| ('convnext_large_d_320', 'laion2b_s29b_b131k_ft_soup')      | 60.91%       | 29.28%       | 38.97%       | 22.61%       | 60.78%       | 65.76%       | 63.84%       | 48.88%     |
| ('convnext_xxlarge', 'laion2b_s34b_b82k_augreg_soup')       | 61.55%       | 30.17%       | 38.85%       | 22.30%       | 60.28%       | 64.83%       | 63.22%       | 48.74%     |
| ('ViT-B-32', 'laion2b_e16')                                 | 61.44%       | 28.15%       | 38.05%       | 24.49%       | 59.93%       | 65.14%       | 63.87%       | 48.72%     |
| ('ViT-B-16', 'datacomp_l_s1b_b8k')                          | 61.33%       | 29.35%       | 38.67%       | 23.31%       | 60.29%       | 64.42%       | 63.64%       | 48.72%     |
| ('ViT-H-14', 'laion2b_s32b_b79k')                           | 61.45%       | 29.19%       | 38.91%       | 22.64%       | 60.56%       | 64.86%       | 63.30%       | 48.70%     |
| ('EVA02-E-14', 'laion2b_s4b_b115k')                         | 61.63%       | 29.60%       | 38.57%       | 22.89%       | 60.22%       | 64.83%       | 63.18%       | 48.70%     |
| ('convnext_xxlarge', 'laion2b_s34b_b82k_augreg_rewind')     | 61.24%       | 30.22%       | 39.04%       | 22.40%       | 60.02%       | 64.75%       | 62.99%       | 48.67%     |
| ('ViT-B-32-quickgelu', 'metaclip_fullcc')                   | 58.26%       | 29.70%       | 38.99%       | 23.24%       | 60.07%       | 65.67%       | 64.30%       | 48.60%     |
| ('convnext_xxlarge', 'laion2b_s34b_b82k_augreg')            | 60.94%       | 29.90%       | 39.49%       | 22.08%       | 60.10%       | 64.50%       | 63.15%       | 48.59%     |
| ('ViT-g-14', 'laion2b_s12b_b42k')                           | 61.46%       | 27.70%       | 38.23%       | 22.46%       | 60.65%       | 65.68%       | 63.87%       | 48.58%     |
| ('ViT-g-14', 'laion2b_s34b_b88k')                           | 60.83%       | 29.56%       | 39.37%       | 21.63%       | 59.87%       | 64.68%       | 63.30%       | 48.46%     |
| ('ViT-L-14-quickgelu', 'metaclip_fullcc')                   | 56.99%       | 31.07%       | 40.45%       | 23.13%       | 59.21%       | 64.77%       | 63.50%       | 48.45%     |
| intfloat/multilingual-e5-small                              | 49.50%       | 42.68%       | 30.96%       | 47.42%       | 54.44%       | 56.44%       | 57.04%       | 48.35%     |
| ('ViT-B-16-quickgelu', 'metaclip_fullcc')                   | 58.00%       | 28.59%       | 37.68%       | 23.22%       | 59.42%       | 65.03%       | 64.10%       | 48.01%     |
| ('ViT-L-14', 'laion2b_s32b_b82k')                           | 60.18%       | 28.09%       | 36.28%       | 23.70%       | 59.89%       | 64.86%       | 63.01%       | 48.00%     |
| ('ViT-B-32-quickgelu', 'laion400m_e32')                     | 59.74%       | 25.92%       | 36.98%       | 25.19%       | 59.67%       | 64.79%       | 63.68%       | 48.00%     |
| ('ViT-B-32-quickgelu', 'laion400m_e31')                     | 59.86%       | 25.92%       | 36.84%       | 25.20%       | 59.56%       | 64.76%       | 63.79%       | 47.99%     |
| ('convnext_base_w', 'laion2b_s13b_b82k_augreg')             | 60.97%       | 27.03%       | 36.75%       | 22.90%       | 59.70%       | 64.78%       | 63.46%       | 47.94%     |
| ('ViT-L-14', 'laion400m_e32')                               | 60.01%       | 24.45%       | 37.24%       | 23.95%       | 59.17%       | 65.02%       | 63.78%       | 47.66%     |
| ('EVA01-g-14', 'laion400m_s11b_b41k')                       | 60.51%       | 25.96%       | 36.17%       | 23.69%       | 59.57%       | 64.40%       | 63.22%       | 47.64%     |
| ('ViT-B-16-plus-240', 'laion400m_e32')                      | 59.84%       | 25.29%       | 36.80%       | 23.73%       | 59.31%       | 64.99%       | 63.43%       | 47.63%     |
| ('ViT-B-16-plus-240', 'laion400m_e31')                      | 59.69%       | 25.22%       | 36.79%       | 23.69%       | 59.44%       | 64.92%       | 63.53%       | 47.61%     |
| ('ViT-B-16', 'laion2b_s34b_b88k')                           | 59.82%       | 27.45%       | 35.12%       | 24.41%       | 59.39%       | 64.37%       | 62.66%       | 47.60%     |
| ('ViT-L-14', 'laion400m_e31')                               | 59.91%       | 24.26%       | 37.53%       | 23.84%       | 59.08%       | 64.90%       | 63.64%       | 47.60%     |
| ('ViT-L-16-SigLIP-256', 'webli')                            | 65.54%       | 20.39%       | 44.65%       | 15.18%       | 60.10%       | 64.64%       | 62.44%       | 47.56%     |
| ('roberta-ViT-B-32', 'laion2b_s12b_b32k')                   | 59.70%       | 25.15%       | 39.81%       | 17.10%       | 59.95%       | 65.81%       | 65.00%       | 47.50%     |
| ('ViT-L-14', 'commonpool_xl_laion_s13b_b90k')               | 58.13%       | 26.95%       | 34.93%       | 23.34%       | 59.05%       | 64.51%       | 63.63%       | 47.22%     |
| ('ViT-B-16-SigLIP', 'webli')                                | 64.31%       | 19.87%       | 44.78%       | 14.87%       | 58.38%       | 65.16%       | 62.44%       | 47.12%     |
| ('ViT-B-16-SigLIP-256', 'webli')                            | 64.24%       | 20.94%       | 44.15%       | 15.35%       | 58.22%       | 64.41%       | 62.43%       | 47.10%     |
| ('ViT-B-16-SigLIP-384', 'webli')                            | 64.36%       | 20.06%       | 44.41%       | 15.11%       | 58.03%       | 64.68%       | 62.10%       | 46.96%     |
| ('ViT-L-16-SigLIP-384', 'webli')                            | 64.49%       | 20.17%       | 44.01%       | 14.80%       | 58.89%       | 64.92%       | 61.39%       | 46.95%     |
| ('ViT-B-32', 'laion400m_e31')                               | 59.06%       | 26.66%       | 35.69%       | 23.68%       | 58.00%       | 62.82%       | 62.68%       | 46.94%     |
| ('ViT-B-16-SigLIP-512', 'webli')                            | 64.28%       | 19.61%       | 44.17%       | 15.09%       | 57.71%       | 64.83%       | 62.44%       | 46.88%     |
| ('convnext_base_w_320', 'laion_aesthetic_s13b_b82k_augreg') | 57.60%       | 26.52%       | 35.01%       | 24.43%       | 57.05%       | 64.54%       | 62.74%       | 46.84%     |
| ('ViT-B-16', 'commonpool_l_text_s1b_b8k')                   | 59.57%       | 28.15%       | 37.37%       | 20.89%       | 57.54%       | 62.68%       | 61.63%       | 46.83%     |
| ('ViT-B-32', 'laion400m_e32')                               | 59.05%       | 26.62%       | 35.44%       | 23.54%       | 58.00%       | 62.74%       | 62.27%       | 46.81%     |
| ('convnext_base_w', 'laion2b_s13b_b82k')                    | 58.65%       | 26.97%       | 34.80%       | 23.26%       | 58.31%       | 63.39%       | 61.56%       | 46.71%     |
| sentence-transformers/gtr-t5-xxl                            | 59.93%       | 24.82%       | 40.79%       | 17.23%       | 58.41%       | 64.00%       | 61.57%       | 46.68%     |
| ('ViT-B-16', 'laion400m_e32')                               | 59.01%       | 24.34%       | 35.07%       | 21.84%       | 59.04%       | 64.58%       | 62.73%       | 46.66%     |
| ('ViT-B-16', 'laion400m_e31')                               | 58.94%       | 24.20%       | 34.92%       | 21.58%       | 59.11%       | 64.77%       | 63.09%       | 46.66%     |
| ('convnext_base', 'laion400m_s13b_b51k')                    | 58.44%       | 24.99%       | 34.05%       | 23.99%       | 58.33%       | 63.79%       | 62.59%       | 46.60%     |
| ('EVA02-L-14-336', 'merged2b_s6b_b61k')                     | 59.54%       | 23.19%       | 34.54%       | 22.36%       | 59.24%       | 63.90%       | 63.40%       | 46.60%     |
| ('coca_ViT-B-32', 'laion2b_s13b_b90k')                      | 58.70%       | 27.10%       | 33.22%       | 24.13%       | 57.53%       | 63.56%       | 61.87%       | 46.59%     |
| ('EVA02-L-14', 'merged2b_s4b_b131k')                        | 59.64%       | 23.18%       | 34.62%       | 22.55%       | 59.11%       | 63.86%       | 63.10%       | 46.58%     |
| thenlper/gte-large                                          | 55.10%       | 28.16%       | 33.96%       | 18.73%       | 59.50%       | 65.19%       | 63.52%       | 46.31%     |
| ('ViT-L-14-quickgelu', 'metaclip_400m')                     | 54.32%       | 25.87%       | 34.30%       | 23.41%       | 58.50%       | 64.48%       | 63.24%       | 46.30%     |
| ('coca_ViT-L-14', 'laion2b_s13b_b90k')                      | 57.92%       | 25.78%       | 33.97%       | 24.17%       | 57.64%       | 63.08%       | 61.55%       | 46.30%     |
| ('coca_ViT-L-14', 'mscoco_finetuned_laion2b_s13b_b90k')     | 58.07%       | 25.32%       | 34.18%       | 24.60%       | 57.77%       | 62.80%       | 61.28%       | 46.29%     |
| ('ViT-B-32-quickgelu', 'metaclip_400m')                     | 55.85%       | 27.37%       | 31.91%       | 21.76%       | 58.64%       | 64.69%       | 63.11%       | 46.19%     |
| sentence-transformers/paraphrase-multilingual-mpnet-base-v2 | 49.03%       | 32.58%       | 32.82%       | 38.43%       | 55.30%       | 57.36%       | 57.34%       | 46.12%     |
| ('convnext_base_w', 'laion_aesthetic_s13b_b82k')            | 57.39%       | 25.68%       | 33.71%       | 23.82%       | 56.64%       | 63.22%       | 62.22%       | 46.10%     |
| ('ViT-B-32', 'commonpool_m_clip_s128m_b4k')                 | 56.09%       | 26.70%       | 38.25%       | 22.79%       | 56.52%       | 61.26%       | 61.05%       | 46.09%     |
| ('convnext_base_w_320', 'laion_aesthetic_s13b_b82k')        | 56.96%       | 25.60%       | 33.77%       | 24.64%       | 56.32%       | 63.33%       | 61.87%       | 46.07%     |
| ('ViT-B-16', 'commonpool_l_laion_s1b_b8k')                  | 56.37%       | 25.70%       | 31.07%       | 23.18%       | 58.65%       | 63.93%       | 63.49%       | 46.06%     |
| ('ViT-B-16-quickgelu', 'metaclip_400m')                     | 55.90%       | 25.88%       | 32.67%       | 21.57%       | 58.65%       | 64.48%       | 63.04%       | 46.03%     |
| intfloat/e5-large                                           | 55.45%       | 28.54%       | 36.69%       | 18.15%       | 57.78%       | 62.92%       | 61.83%       | 45.91%     |
| ('EVA02-B-16', 'merged2b_s8b_b131k')                        | 58.08%       | 24.45%       | 31.80%       | 22.36%       | 58.45%       | 63.25%       | 62.44%       | 45.83%     |
| sentence-transformers/LaBSE                                 | 50.30%       | 32.82%       | 33.15%       | 39.79%       | 54.95%       | 53.71%       | 55.06%       | 45.68%     |
| thenlper/gte-base                                           | 55.46%       | 27.88%       | 32.77%       | 17.20%       | 58.09%       | 63.68%       | 62.03%       | 45.30%     |
| intfloat/e5-large-v2                                        | 55.10%       | 28.06%       | 35.95%       | 17.16%       | 57.16%       | 61.21%       | 60.84%       | 45.07%     |
| ('ViT-SO400M-14-SigLIP', 'webli')                           | 60.18%       | 29.39%       | 38.90%       | 13.73%       | 52.79%       | 59.15%       | 56.81%       | 44.42%     |
| ('ViT-B-32', 'commonpool_m_s128m_b4k')                      | 50.30%       | 32.12%       | 37.08%       | 23.02%       | 53.63%       | 57.64%       | 56.91%       | 44.39%     |
| sentence-transformers/sentence-t5-xxl                       | 50.98%       | 18.38%       | 36.37%       | 16.91%       | 59.25%       | 64.82%       | 63.75%       | 44.35%     |
| infgrad/stella-base-en-v2                                   | 52.42%       | 26.24%       | 30.61%       | 18.81%       | 56.84%       | 63.03%       | 61.67%       | 44.23%     |
| ('RN50x4', 'openai')                                        | 56.39%       | 25.77%       | 29.99%       | 21.48%       | 55.31%       | 61.02%       | 59.42%       | 44.20%     |
| ('RN50x16', 'openai')                                       | 56.58%       | 25.09%       | 29.77%       | 21.03%       | 54.81%       | 61.28%       | 58.47%       | 43.86%     |
| ('RN101-quickgelu', 'openai')                               | 56.57%       | 25.83%       | 29.66%       | 21.09%       | 54.50%       | 60.18%       | 58.74%       | 43.80%     |
| ('RN101', 'openai')                                         | 56.57%       | 25.83%       | 29.66%       | 21.09%       | 54.50%       | 60.18%       | 58.74%       | 43.80%     |
| llmrails/ember-v1                                           | 50.85%       | 24.76%       | 31.02%       | 17.20%       | 57.62%       | 63.06%       | 62.04%       | 43.79%     |
| sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 | 44.88%       | 28.32%       | 29.45%       | 36.40%       | 53.97%       | 56.87%       | 56.14%       | 43.72%     |
| BAAI/bge-large-en-v1.5                                      | 49.81%       | 25.55%       | 30.68%       | 17.41%       | 56.89%       | 62.87%       | 61.72%       | 43.56%     |
| ('RN50x64', 'openai')                                       | 55.34%       | 22.19%       | 30.63%       | 20.79%       | 55.18%       | 60.93%       | 59.45%       | 43.50%     |
| ('nllb-clip-large', 'v1')                                   | 48.84%       | 23.45%       | 33.92%       | 32.38%       | 53.67%       | 55.36%       | 56.76%       | 43.48%     |
| BAAI/bge-base-en-v1.5                                       | 51.73%       | 24.30%       | 31.51%       | 17.53%       | 56.21%       | 62.37%       | 60.25%       | 43.42%     |
| intfloat/e5-small                                           | 51.31%       | 27.36%       | 32.05%       | 16.66%       | 55.15%       | 60.39%       | 59.06%       | 43.14%     |
| BAAI/bge-small-en-v1.5                                      | 51.37%       | 25.16%       | 29.99%       | 16.13%       | 56.17%       | 61.69%       | 61.01%       | 43.07%     |
| ('ViT-L-14', 'openai')                                      | 54.57%       | 21.44%       | 30.13%       | 19.50%       | 54.99%       | 60.94%       | 59.59%       | 43.02%     |
| ('ViT-L-14-336', 'openai')                                  | 54.12%       | 21.52%       | 30.63%       | 19.47%       | 55.41%       | 60.77%       | 58.87%       | 42.97%     |
| intfloat/e5-small-v2                                        | 51.41%       | 26.82%       | 33.04%       | 16.30%       | 54.97%       | 58.66%       | 58.68%       | 42.84%     |
| ('ViT-SO400M-14-SigLIP-384', 'webli')                       | 62.68%       | 15.00%       | 32.38%       | 7.32%        | 56.65%       | 64.12%       | 61.49%       | 42.81%     |
| ('RN50-quickgelu', 'openai')                                | 53.15%       | 24.79%       | 29.57%       | 20.84%       | 53.15%       | 59.19%       | 57.59%       | 42.61%     |
| ('RN50', 'openai')                                          | 53.15%       | 24.79%       | 29.57%       | 20.84%       | 53.15%       | 59.19%       | 57.59%       | 42.61%     |
| ('ViT-B-16', 'openai')                                      | 53.31%       | 22.22%       | 27.96%       | 21.22%       | 53.68%       | 59.47%       | 58.45%       | 42.33%     |
| ('ViT-B-32', 'openai')                                      | 52.93%       | 23.44%       | 28.70%       | 20.78%       | 52.96%       | 59.38%       | 57.93%       | 42.30%     |
| ('ViT-B-32-quickgelu', 'openai')                            | 52.93%       | 23.44%       | 28.70%       | 20.78%       | 52.96%       | 59.38%       | 57.93%       | 42.30%     |
| sentence-transformers/all-MiniLM-L6-v2                      | 50.80%       | 25.76%       | 27.04%       | 15.81%       | 54.63%       | 60.07%       | 59.68%       | 41.97%     |
| ('ViT-B-32', 'commonpool_m_basic_s128m_b4k')                | 52.54%       | 22.67%       | 30.25%       | 16.17%       | 53.22%       | 59.40%       | 58.31%       | 41.80%     |
| sentence-transformers/all-MiniLM-L12-v2                     | 48.98%       | 24.05%       | 25.74%       | 16.41%       | 54.51%       | 60.38%       | 58.90%       | 41.28%     |
| ('ViT-B-32', 'commonpool_m_image_s128m_b4k')                | 51.93%       | 20.40%       | 29.44%       | 16.53%       | 53.16%       | 58.71%       | 58.17%       | 41.19%     |
| sentence-transformers/clip-ViT-B-32-multilingual-v1         | 44.45%       | 27.34%       | 28.00%       | 28.25%       | 50.30%       | 54.05%       | 53.39%       | 40.82%     |
| sentence-transformers/distiluse-base-multilingual-cased-v2  | 43.51%       | 23.86%       | 28.41%       | 26.90%       | 53.14%       | 53.54%       | 54.38%       | 40.53%     |
| ('ViT-B-32', 'datacomp_m_s128m_b4k')                        | 51.60%       | 19.45%       | 26.58%       | 16.46%       | 52.54%       | 59.03%       | 58.03%       | 40.53%     |
| ('ViT-B-32', 'commonpool_m_text_s128m_b4k')                 | 50.38%       | 20.31%       | 27.01%       | 16.00%       | 52.61%       | 58.82%       | 58.10%       | 40.46%     |
| sentence-transformers/all-mpnet-base-v2                     | 46.97%       | 23.15%       | 24.75%       | 16.31%       | 52.66%       | 59.07%       | 57.75%       | 40.09%     |
| ('nllb-clip-base', 'v1')                                    | 42.72%       | 23.90%       | 29.29%       | 33.96%       | 48.33%       | 49.09%       | 51.21%       | 39.79%     |
| sentence-transformers/paraphrase-mpnet-base-v2              | 46.00%       | 20.45%       | 26.92%       | 14.75%       | 52.89%       | 58.71%       | 58.20%       | 39.70%     |
| sentence-transformers/all-distilroberta-v1                  | 46.74%       | 22.34%       | 24.06%       | 17.59%       | 51.49%       | 57.54%       | 56.45%       | 39.46%     |
| sentence-transformers/paraphrase-MiniLM-L6-v2               | 44.92%       | 23.59%       | 26.12%       | 14.23%       | 51.84%       | 57.14%       | 56.03%       | 39.12%     |
| ('ViT-B-32', 'commonpool_m_laion_s128m_b4k')                | 42.94%       | 19.21%       | 19.70%       | 17.26%       | 50.84%       | 57.59%       | 56.06%       | 37.66%     |
| ('RN50-quickgelu', 'cc12m')                                 | 40.71%       | 18.10%       | 16.78%       | 16.23%       | 45.55%       | 52.89%       | 50.77%       | 34.43%     |
| ('RN50', 'cc12m')                                           | 39.76%       | 17.32%       | 16.15%       | 15.76%       | 44.25%       | 52.46%       | 49.18%       | 33.55%     |
| ('RN101', 'yfcc15m')                                        | 33.79%       | 18.04%       | 16.05%       | 11.10%       | 37.62%       | 43.50%       | 42.45%       | 28.94%     |
| ('RN101-quickgelu', 'yfcc15m')                              | 32.79%       | 16.89%       | 14.45%       | 11.56%       | 37.77%       | 42.86%       | 41.93%       | 28.32%     |
| ('ViT-B-32', 'commonpool_s_clip_s13m_b4k')                  | 33.80%       | 13.26%       | 18.82%       | 12.42%       | 37.36%       | 42.09%       | 40.39%       | 28.31%     |
| ('RN50', 'yfcc15m')                                         | 31.81%       | 15.87%       | 14.88%       | 8.99%        | 37.42%       | 42.06%       | 41.19%       | 27.46%     |
| ('RN50-quickgelu', 'yfcc15m')                               | 31.57%       | 15.90%       | 14.44%       | 8.99%        | 36.81%       | 41.81%       | 41.20%       | 27.24%     |
| ('ViT-B-32', 'commonpool_s_s13m_b4k')                       | 29.42%       | 12.57%       | 16.82%       | 11.00%       | 32.42%       | 36.77%       | 35.48%       | 24.93%     |
| ('ViT-B-32', 'commonpool_s_text_s13m_b4k')                  | 28.02%       | 10.61%       | 12.49%       | 9.85%        | 31.18%       | 37.10%       | 34.85%       | 23.44%     |
| ('ViT-B-32', 'commonpool_s_basic_s13m_b4k')                 | 27.87%       | 10.72%       | 12.67%       | 8.16%        | 30.11%       | 36.13%       | 32.68%       | 22.62%     |
| ('coca_ViT-B-32', 'mscoco_finetuned_laion2b_s13b_b90k')     | 12.60%       | 7.91%        | 5.11%        | 9.96%        | 17.15%       | 20.67%       | 20.32%       | 13.39%     |
| ('ViT-B-32', 'commonpool_s_image_s13m_b4k')                 | 15.20%       | 5.59%        | 5.91%        | 4.63%        | 16.80%       | 20.74%       | 18.78%       | 12.52%     |
| ('ViT-B-32', 'datacomp_s_s13m_b4k')                         | 15.20%       | 5.59%        | 5.91%        | 4.63%        | 16.80%       | 20.74%       | 18.78%       | 12.52%     |
| ('ViT-B-32', 'commonpool_s_laion_s13m_b4k')                 | 11.72%       | 5.12%        | 4.05%        | 4.23%        | 14.33%       | 18.82%       | 16.44%       | 10.67%     |

## Training
The model was trained with the parameters:

**DataLoader**:

`torch.utils.data.dataloader.DataLoader` of length 1468721 with parameters:
```
{'batch_size': 160, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
```

**Loss**:

`sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` 

Parameters of the fit()-Method:
```
{
    "epochs": 1,
    "evaluation_steps": 0,
    "evaluator": "NoneType",
    "max_grad_norm": 1,
    "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
    "optimizer_params": {
        "lr": 2e-05
    },
    "scheduler": "WarmupLinear",
    "steps_per_epoch": null,
    "warmup_steps": 100,
    "weight_decay": 0.01
}
```


## Full Model Architecture
```
SentenceTransformer(
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel 
  (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
```

## Citing & Authors

<!--- Describe where people can find more information -->