Datasets:
mteb
/

Sam commited on
Commit
f0b35fd
·
unverified ·
1 Parent(s): 7573d55

add german results for ger models (#96)

Browse files
Files changed (26) hide show
  1. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/AmazonReviewsClassification.json +137 -0
  2. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/BlurbsClusteringP2P.json +50 -0
  3. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/BlurbsClusteringS2S.json +50 -0
  4. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MTOPDomainClassification.json +137 -0
  5. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MTOPIntentClassification.json +137 -0
  6. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MassiveScenarioClassification.json +196 -72
  7. results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/model_meta.json +1 -1
  8. results/Cohere__Cohere-embed-multilingual-light-v3.0/1/model_meta.json +1 -1
  9. results/Cohere__Cohere-embed-multilingual-v3.0/1/model_meta.json +1 -1
  10. results/Snowflake__snowflake-arctic-embed-l-v2.0/edc2df7b6c25794b340229ca082e7c78782e6374/GermanDPR.json +158 -0
  11. results/Snowflake__snowflake-arctic-embed-l-v2.0/edc2df7b6c25794b340229ca082e7c78782e6374/model_meta.json +1 -1
  12. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/AmazonCounterfactualClassification.json +181 -0
  13. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/AmazonReviewsClassification.json +137 -0
  14. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/FalseFriendsGermanEnglish.json +58 -0
  15. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/GermanDPR.json +158 -0
  16. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/GermanQuAD-Retrieval.json +158 -0
  17. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/MTOPDomainClassification.json +137 -0
  18. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/MTOPIntentClassification.json +137 -0
  19. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/PawsXPairClassification.json +107 -0
  20. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/XMarket.json +158 -0
  21. results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/model_meta.json +1 -1
  22. results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/AmazonCounterfactualClassification.json +181 -0
  23. results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/AmazonReviewsClassification.json +137 -0
  24. results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/MTOPDomainClassification.json +137 -0
  25. results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/MTOPIntentClassification.json +137 -0
  26. results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/model_meta.json +1 -1
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/AmazonReviewsClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1399c76144fd37290681b995c656ef9b2e06e26d",
3
+ "task_name": "AmazonReviewsClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.4614,
9
+ "f1": 0.435978,
10
+ "f1_weighted": 0.435978,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.4848,
14
+ "f1": 0.46934,
15
+ "f1_weighted": 0.46934
16
+ },
17
+ {
18
+ "accuracy": 0.4628,
19
+ "f1": 0.43203,
20
+ "f1_weighted": 0.43203
21
+ },
22
+ {
23
+ "accuracy": 0.4528,
24
+ "f1": 0.421687,
25
+ "f1_weighted": 0.421687
26
+ },
27
+ {
28
+ "accuracy": 0.4986,
29
+ "f1": 0.467483,
30
+ "f1_weighted": 0.467483
31
+ },
32
+ {
33
+ "accuracy": 0.471,
34
+ "f1": 0.443669,
35
+ "f1_weighted": 0.443669
36
+ },
37
+ {
38
+ "accuracy": 0.464,
39
+ "f1": 0.432224,
40
+ "f1_weighted": 0.432224
41
+ },
42
+ {
43
+ "accuracy": 0.4034,
44
+ "f1": 0.396426,
45
+ "f1_weighted": 0.396426
46
+ },
47
+ {
48
+ "accuracy": 0.4938,
49
+ "f1": 0.466621,
50
+ "f1_weighted": 0.466621
51
+ },
52
+ {
53
+ "accuracy": 0.4492,
54
+ "f1": 0.42017,
55
+ "f1_weighted": 0.42017
56
+ },
57
+ {
58
+ "accuracy": 0.4336,
59
+ "f1": 0.410134,
60
+ "f1_weighted": 0.410134
61
+ }
62
+ ],
63
+ "main_score": 0.4614,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.46368,
73
+ "f1": 0.439507,
74
+ "f1_weighted": 0.439507,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.4806,
78
+ "f1": 0.463278,
79
+ "f1_weighted": 0.463278
80
+ },
81
+ {
82
+ "accuracy": 0.4642,
83
+ "f1": 0.433653,
84
+ "f1_weighted": 0.433653
85
+ },
86
+ {
87
+ "accuracy": 0.4534,
88
+ "f1": 0.423666,
89
+ "f1_weighted": 0.423666
90
+ },
91
+ {
92
+ "accuracy": 0.4966,
93
+ "f1": 0.468083,
94
+ "f1_weighted": 0.468083
95
+ },
96
+ {
97
+ "accuracy": 0.4776,
98
+ "f1": 0.450634,
99
+ "f1_weighted": 0.450634
100
+ },
101
+ {
102
+ "accuracy": 0.4768,
103
+ "f1": 0.450572,
104
+ "f1_weighted": 0.450572
105
+ },
106
+ {
107
+ "accuracy": 0.4174,
108
+ "f1": 0.409543,
109
+ "f1_weighted": 0.409543
110
+ },
111
+ {
112
+ "accuracy": 0.4944,
113
+ "f1": 0.46879,
114
+ "f1_weighted": 0.46879
115
+ },
116
+ {
117
+ "accuracy": 0.4498,
118
+ "f1": 0.419052,
119
+ "f1_weighted": 0.419052
120
+ },
121
+ {
122
+ "accuracy": 0.426,
123
+ "f1": 0.407799,
124
+ "f1_weighted": 0.407799
125
+ }
126
+ ],
127
+ "main_score": 0.46368,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 126.27368593215942,
136
+ "kg_co2_emissions": null
137
+ }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/BlurbsClusteringP2P.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "a2dd5b02a77de3466a3eaa98ae586b5610314496",
3
+ "task_name": "BlurbsClusteringP2P",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measure": 0.386876,
9
+ "v_measure_std": 0.083317,
10
+ "v_measures": [
11
+ 0.336807,
12
+ 0.321236,
13
+ 0.312527,
14
+ 0.321698,
15
+ 0.315842,
16
+ 0.318106,
17
+ 0.344385,
18
+ 0.303119,
19
+ 0.327119,
20
+ 0.324908,
21
+ 0.366773,
22
+ 0.496408,
23
+ 0.334427,
24
+ 0.286325,
25
+ 0.264392,
26
+ 0.364911,
27
+ 0.24358,
28
+ 0.495096,
29
+ 0.479464,
30
+ 0.479456,
31
+ 0.475731,
32
+ 0.472335,
33
+ 0.475948,
34
+ 0.461984,
35
+ 0.480143,
36
+ 0.478474,
37
+ 0.475722,
38
+ 0.475621
39
+ ],
40
+ "main_score": 0.386876,
41
+ "hf_subset": "default",
42
+ "languages": [
43
+ "deu-Latn"
44
+ ]
45
+ }
46
+ ]
47
+ },
48
+ "evaluation_time": 4207.973029613495,
49
+ "kg_co2_emissions": null
50
+ }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/BlurbsClusteringS2S.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "22793b6a6465bf00120ad525e38c51210858132c",
3
+ "task_name": "BlurbsClusteringS2S",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measure": 0.162811,
9
+ "v_measure_std": 0.082541,
10
+ "v_measures": [
11
+ 0.091455,
12
+ 0.077184,
13
+ 0.106489,
14
+ 0.087398,
15
+ 0.07582,
16
+ 0.100122,
17
+ 0.088196,
18
+ 0.077762,
19
+ 0.100258,
20
+ 0.094399,
21
+ 0.124095,
22
+ 0.246679,
23
+ 0.051443,
24
+ 0.085061,
25
+ 0.094462,
26
+ 0.223444,
27
+ 0.100991,
28
+ 0.322938,
29
+ 0.217459,
30
+ 0.208868,
31
+ 0.206221,
32
+ 0.214174,
33
+ 0.271484,
34
+ 0.218194,
35
+ 0.230808,
36
+ 0.297791,
37
+ 0.284434,
38
+ 0.261075
39
+ ],
40
+ "main_score": 0.162811,
41
+ "hf_subset": "default",
42
+ "languages": [
43
+ "deu-Latn"
44
+ ]
45
+ }
46
+ ]
47
+ },
48
+ "evaluation_time": 278.6751754283905,
49
+ "kg_co2_emissions": null
50
+ }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MTOPDomainClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d80d48c1eb48d3562165c59d59d0034df9fff0bf",
3
+ "task_name": "MTOPDomainClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.904848,
9
+ "f1": 0.896445,
10
+ "f1_weighted": 0.904566,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.888705,
14
+ "f1": 0.883251,
15
+ "f1_weighted": 0.888924
16
+ },
17
+ {
18
+ "accuracy": 0.905234,
19
+ "f1": 0.894371,
20
+ "f1_weighted": 0.90487
21
+ },
22
+ {
23
+ "accuracy": 0.907989,
24
+ "f1": 0.899887,
25
+ "f1_weighted": 0.907398
26
+ },
27
+ {
28
+ "accuracy": 0.912397,
29
+ "f1": 0.904507,
30
+ "f1_weighted": 0.911696
31
+ },
32
+ {
33
+ "accuracy": 0.905785,
34
+ "f1": 0.895272,
35
+ "f1_weighted": 0.904943
36
+ },
37
+ {
38
+ "accuracy": 0.906887,
39
+ "f1": 0.895359,
40
+ "f1_weighted": 0.90575
41
+ },
42
+ {
43
+ "accuracy": 0.905785,
44
+ "f1": 0.897953,
45
+ "f1_weighted": 0.905034
46
+ },
47
+ {
48
+ "accuracy": 0.898072,
49
+ "f1": 0.894061,
50
+ "f1_weighted": 0.899223
51
+ },
52
+ {
53
+ "accuracy": 0.902479,
54
+ "f1": 0.894367,
55
+ "f1_weighted": 0.902354
56
+ },
57
+ {
58
+ "accuracy": 0.915152,
59
+ "f1": 0.905425,
60
+ "f1_weighted": 0.915466
61
+ }
62
+ ],
63
+ "main_score": 0.904848,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.912229,
73
+ "f1": 0.902667,
74
+ "f1_weighted": 0.911946,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.899408,
78
+ "f1": 0.892616,
79
+ "f1_weighted": 0.899651
80
+ },
81
+ {
82
+ "accuracy": 0.911524,
83
+ "f1": 0.901484,
84
+ "f1_weighted": 0.910879
85
+ },
86
+ {
87
+ "accuracy": 0.915187,
88
+ "f1": 0.904778,
89
+ "f1_weighted": 0.914295
90
+ },
91
+ {
92
+ "accuracy": 0.912651,
93
+ "f1": 0.90301,
94
+ "f1_weighted": 0.911949
95
+ },
96
+ {
97
+ "accuracy": 0.917723,
98
+ "f1": 0.905485,
99
+ "f1_weighted": 0.91666
100
+ },
101
+ {
102
+ "accuracy": 0.912933,
103
+ "f1": 0.900559,
104
+ "f1_weighted": 0.912304
105
+ },
106
+ {
107
+ "accuracy": 0.918287,
108
+ "f1": 0.907105,
109
+ "f1_weighted": 0.917834
110
+ },
111
+ {
112
+ "accuracy": 0.897999,
113
+ "f1": 0.891261,
114
+ "f1_weighted": 0.899309
115
+ },
116
+ {
117
+ "accuracy": 0.913215,
118
+ "f1": 0.90552,
119
+ "f1_weighted": 0.912925
120
+ },
121
+ {
122
+ "accuracy": 0.923359,
123
+ "f1": 0.914857,
124
+ "f1_weighted": 0.923654
125
+ }
126
+ ],
127
+ "main_score": 0.912229,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 17.621785640716553,
136
+ "kg_co2_emissions": null
137
+ }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MTOPIntentClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba",
3
+ "task_name": "MTOPIntentClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.657741,
9
+ "f1": 0.433268,
10
+ "f1_weighted": 0.690047,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.661157,
14
+ "f1": 0.432011,
15
+ "f1_weighted": 0.696916
16
+ },
17
+ {
18
+ "accuracy": 0.678788,
19
+ "f1": 0.466991,
20
+ "f1_weighted": 0.714941
21
+ },
22
+ {
23
+ "accuracy": 0.663361,
24
+ "f1": 0.43114,
25
+ "f1_weighted": 0.696194
26
+ },
27
+ {
28
+ "accuracy": 0.649587,
29
+ "f1": 0.432822,
30
+ "f1_weighted": 0.674071
31
+ },
32
+ {
33
+ "accuracy": 0.662259,
34
+ "f1": 0.432395,
35
+ "f1_weighted": 0.69183
36
+ },
37
+ {
38
+ "accuracy": 0.627548,
39
+ "f1": 0.427565,
40
+ "f1_weighted": 0.660601
41
+ },
42
+ {
43
+ "accuracy": 0.672727,
44
+ "f1": 0.431111,
45
+ "f1_weighted": 0.700139
46
+ },
47
+ {
48
+ "accuracy": 0.660606,
49
+ "f1": 0.442187,
50
+ "f1_weighted": 0.696946
51
+ },
52
+ {
53
+ "accuracy": 0.650689,
54
+ "f1": 0.420515,
55
+ "f1_weighted": 0.684475
56
+ },
57
+ {
58
+ "accuracy": 0.650689,
59
+ "f1": 0.415943,
60
+ "f1_weighted": 0.684352
61
+ }
62
+ ],
63
+ "main_score": 0.657741,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.68084,
73
+ "f1": 0.461842,
74
+ "f1_weighted": 0.711913,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.686391,
78
+ "f1": 0.470712,
79
+ "f1_weighted": 0.717191
80
+ },
81
+ {
82
+ "accuracy": 0.694844,
83
+ "f1": 0.472763,
84
+ "f1_weighted": 0.731372
85
+ },
86
+ {
87
+ "accuracy": 0.686954,
88
+ "f1": 0.459152,
89
+ "f1_weighted": 0.720311
90
+ },
91
+ {
92
+ "accuracy": 0.674556,
93
+ "f1": 0.468172,
94
+ "f1_weighted": 0.698181
95
+ },
96
+ {
97
+ "accuracy": 0.684982,
98
+ "f1": 0.464202,
99
+ "f1_weighted": 0.7154
100
+ },
101
+ {
102
+ "accuracy": 0.64328,
103
+ "f1": 0.44399,
104
+ "f1_weighted": 0.674767
105
+ },
106
+ {
107
+ "accuracy": 0.691181,
108
+ "f1": 0.460218,
109
+ "f1_weighted": 0.717318
110
+ },
111
+ {
112
+ "accuracy": 0.681037,
113
+ "f1": 0.467666,
114
+ "f1_weighted": 0.714447
115
+ },
116
+ {
117
+ "accuracy": 0.680473,
118
+ "f1": 0.453633,
119
+ "f1_weighted": 0.713016
120
+ },
121
+ {
122
+ "accuracy": 0.6847,
123
+ "f1": 0.457912,
124
+ "f1_weighted": 0.717126
125
+ }
126
+ ],
127
+ "main_score": 0.68084,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 51.36896634101868,
136
+ "kg_co2_emissions": null
137
+ }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/MassiveScenarioClassification.json CHANGED
@@ -1,137 +1,261 @@
1
  {
2
  "dataset_revision": "fad2c6e8459f9e1c45d9315f4953d921437d70f8",
3
- "evaluation_time": 13.257129430770874,
4
- "kg_co2_emissions": 0.0024129392078880794,
5
- "mteb_version": "1.14.12",
6
  "scores": {
7
  "test": [
8
  {
9
- "accuracy": 0.7342299932750505,
10
- "f1": 0.727081073738889,
11
- "f1_weighted": 0.7291454381904577,
12
  "hf_subset": "ru",
13
  "languages": [
14
  "rus-Cyrl"
15
  ],
16
- "main_score": 0.7342299932750505,
17
  "scores_per_experiment": [
18
  {
19
- "accuracy": 0.7370544720914594,
20
- "f1": 0.7320001869140739,
21
- "f1_weighted": 0.7322000402888483
22
  },
23
  {
24
- "accuracy": 0.7320107599193006,
25
- "f1": 0.7312433708999367,
26
- "f1_weighted": 0.7266766891375795
27
  },
28
  {
29
- "accuracy": 0.7437794216543376,
30
- "f1": 0.7305428360973807,
31
- "f1_weighted": 0.7393301741691624
32
  },
33
  {
34
- "accuracy": 0.7424344317417619,
35
- "f1": 0.7358147374978421,
36
- "f1_weighted": 0.7392674329786143
37
  },
38
  {
39
- "accuracy": 0.7387357094821789,
40
- "f1": 0.7239180276829599,
41
- "f1_weighted": 0.7295326553567506
42
  },
43
  {
44
- "accuracy": 0.7081371889710827,
45
- "f1": 0.6974759366496573,
46
- "f1_weighted": 0.6999339182674899
47
  },
48
  {
49
- "accuracy": 0.7310020174848688,
50
- "f1": 0.7223636587611565,
51
- "f1_weighted": 0.7270063978481861
52
  },
53
  {
54
- "accuracy": 0.7135171486213854,
55
- "f1": 0.7135032404031129,
56
- "f1_weighted": 0.7091690336631582
57
  },
58
  {
59
- "accuracy": 0.7508406186953598,
60
- "f1": 0.7452560010429398,
61
- "f1_weighted": 0.7487309951111808
62
  },
63
  {
64
- "accuracy": 0.7447881640887694,
65
- "f1": 0.7386927414398298,
66
- "f1_weighted": 0.7396070450836079
67
  }
68
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  }
70
  ],
71
  "validation": [
72
  {
73
- "accuracy": 0.7334481062469258,
74
- "f1": 0.7219642986686683,
75
- "f1_weighted": 0.7291686660772714,
76
  "hf_subset": "ru",
77
  "languages": [
78
  "rus-Cyrl"
79
  ],
80
- "main_score": 0.7334481062469258,
81
  "scores_per_experiment": [
82
  {
83
- "accuracy": 0.750614854894245,
84
- "f1": 0.7419129503258692,
85
- "f1_weighted": 0.7482714805119889
86
  },
87
  {
88
- "accuracy": 0.721593703885883,
89
- "f1": 0.7173307078578347,
90
- "f1_weighted": 0.7184545272456723
91
  },
92
  {
93
- "accuracy": 0.7545499262174127,
94
- "f1": 0.7386459559639673,
95
- "f1_weighted": 0.7526213558774643
96
  },
97
  {
98
- "accuracy": 0.7220855878012788,
99
- "f1": 0.7137947636981896,
100
- "f1_weighted": 0.7165408498996397
101
  },
102
  {
103
- "accuracy": 0.7437284800787014,
104
- "f1": 0.7227735633917522,
105
- "f1_weighted": 0.7362798488981783
106
  },
107
  {
108
- "accuracy": 0.6960157402852927,
109
- "f1": 0.6873383424519843,
110
- "f1_weighted": 0.6875594902710235
111
  },
112
  {
113
- "accuracy": 0.7122479094933596,
114
- "f1": 0.6940667638934372,
115
- "f1_weighted": 0.7083866754061244
116
  },
117
  {
118
- "accuracy": 0.7255287752090507,
119
- "f1": 0.7176890888958867,
120
- "f1_weighted": 0.7231797626602348
121
  },
122
  {
123
- "accuracy": 0.7594687653713723,
124
- "f1": 0.7506790383048325,
125
- "f1_weighted": 0.7563092326920408
126
  },
127
  {
128
- "accuracy": 0.7486473192326611,
129
- "f1": 0.7354118119029295,
130
- "f1_weighted": 0.7440834373103473
131
  }
132
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133
  }
134
  ]
135
  },
136
- "task_name": "MassiveScenarioClassification"
 
137
  }
 
1
  {
2
  "dataset_revision": "fad2c6e8459f9e1c45d9315f4953d921437d70f8",
3
+ "task_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.29.10",
 
5
  "scores": {
6
  "test": [
7
  {
8
+ "accuracy": 0.73423,
9
+ "f1": 0.727081,
10
+ "f1_weighted": 0.729145,
11
  "hf_subset": "ru",
12
  "languages": [
13
  "rus-Cyrl"
14
  ],
15
+ "main_score": 0.73423,
16
  "scores_per_experiment": [
17
  {
18
+ "accuracy": 0.737054,
19
+ "f1": 0.732,
20
+ "f1_weighted": 0.7322
21
  },
22
  {
23
+ "accuracy": 0.732011,
24
+ "f1": 0.731243,
25
+ "f1_weighted": 0.726677
26
  },
27
  {
28
+ "accuracy": 0.743779,
29
+ "f1": 0.730543,
30
+ "f1_weighted": 0.73933
31
  },
32
  {
33
+ "accuracy": 0.742434,
34
+ "f1": 0.735815,
35
+ "f1_weighted": 0.739267
36
  },
37
  {
38
+ "accuracy": 0.738736,
39
+ "f1": 0.723918,
40
+ "f1_weighted": 0.729533
41
  },
42
  {
43
+ "accuracy": 0.708137,
44
+ "f1": 0.697476,
45
+ "f1_weighted": 0.699934
46
  },
47
  {
48
+ "accuracy": 0.731002,
49
+ "f1": 0.722364,
50
+ "f1_weighted": 0.727006
51
  },
52
  {
53
+ "accuracy": 0.713517,
54
+ "f1": 0.713503,
55
+ "f1_weighted": 0.709169
56
  },
57
  {
58
+ "accuracy": 0.750841,
59
+ "f1": 0.745256,
60
+ "f1_weighted": 0.748731
61
  },
62
  {
63
+ "accuracy": 0.744788,
64
+ "f1": 0.738693,
65
+ "f1_weighted": 0.739607
66
  }
67
  ]
68
+ },
69
+ {
70
+ "accuracy": 0.73544,
71
+ "f1": 0.727401,
72
+ "f1_weighted": 0.728997,
73
+ "scores_per_experiment": [
74
+ {
75
+ "accuracy": 0.742098,
76
+ "f1": 0.733146,
77
+ "f1_weighted": 0.737209
78
+ },
79
+ {
80
+ "accuracy": 0.747478,
81
+ "f1": 0.740965,
82
+ "f1_weighted": 0.741591
83
+ },
84
+ {
85
+ "accuracy": 0.735037,
86
+ "f1": 0.727365,
87
+ "f1_weighted": 0.728237
88
+ },
89
+ {
90
+ "accuracy": 0.731338,
91
+ "f1": 0.72088,
92
+ "f1_weighted": 0.727566
93
+ },
94
+ {
95
+ "accuracy": 0.732683,
96
+ "f1": 0.726029,
97
+ "f1_weighted": 0.722856
98
+ },
99
+ {
100
+ "accuracy": 0.70881,
101
+ "f1": 0.69902,
102
+ "f1_weighted": 0.69472
103
+ },
104
+ {
105
+ "accuracy": 0.736718,
106
+ "f1": 0.724222,
107
+ "f1_weighted": 0.728796
108
+ },
109
+ {
110
+ "accuracy": 0.725286,
111
+ "f1": 0.724655,
112
+ "f1_weighted": 0.724493
113
+ },
114
+ {
115
+ "accuracy": 0.750841,
116
+ "f1": 0.744536,
117
+ "f1_weighted": 0.747324
118
+ },
119
+ {
120
+ "accuracy": 0.744116,
121
+ "f1": 0.733189,
122
+ "f1_weighted": 0.737177
123
+ }
124
+ ],
125
+ "main_score": 0.73544,
126
+ "hf_subset": "de",
127
+ "languages": [
128
+ "deu-Latn"
129
+ ]
130
  }
131
  ],
132
  "validation": [
133
  {
134
+ "accuracy": 0.733448,
135
+ "f1": 0.721964,
136
+ "f1_weighted": 0.729169,
137
  "hf_subset": "ru",
138
  "languages": [
139
  "rus-Cyrl"
140
  ],
141
+ "main_score": 0.733448,
142
  "scores_per_experiment": [
143
  {
144
+ "accuracy": 0.750615,
145
+ "f1": 0.741913,
146
+ "f1_weighted": 0.748271
147
  },
148
  {
149
+ "accuracy": 0.721594,
150
+ "f1": 0.717331,
151
+ "f1_weighted": 0.718455
152
  },
153
  {
154
+ "accuracy": 0.75455,
155
+ "f1": 0.738646,
156
+ "f1_weighted": 0.752621
157
  },
158
  {
159
+ "accuracy": 0.722086,
160
+ "f1": 0.713795,
161
+ "f1_weighted": 0.716541
162
  },
163
  {
164
+ "accuracy": 0.743728,
165
+ "f1": 0.722774,
166
+ "f1_weighted": 0.73628
167
  },
168
  {
169
+ "accuracy": 0.696016,
170
+ "f1": 0.687338,
171
+ "f1_weighted": 0.687559
172
  },
173
  {
174
+ "accuracy": 0.712248,
175
+ "f1": 0.694067,
176
+ "f1_weighted": 0.708387
177
  },
178
  {
179
+ "accuracy": 0.725529,
180
+ "f1": 0.717689,
181
+ "f1_weighted": 0.72318
182
  },
183
  {
184
+ "accuracy": 0.759469,
185
+ "f1": 0.750679,
186
+ "f1_weighted": 0.756309
187
  },
188
  {
189
+ "accuracy": 0.748647,
190
+ "f1": 0.735412,
191
+ "f1_weighted": 0.744083
192
  }
193
  ]
194
+ },
195
+ {
196
+ "accuracy": 0.728037,
197
+ "f1": 0.714535,
198
+ "f1_weighted": 0.722801,
199
+ "scores_per_experiment": [
200
+ {
201
+ "accuracy": 0.740285,
202
+ "f1": 0.72853,
203
+ "f1_weighted": 0.736101
204
+ },
205
+ {
206
+ "accuracy": 0.739302,
207
+ "f1": 0.725693,
208
+ "f1_weighted": 0.737135
209
+ },
210
+ {
211
+ "accuracy": 0.727004,
212
+ "f1": 0.715785,
213
+ "f1_weighted": 0.72249
214
+ },
215
+ {
216
+ "accuracy": 0.722086,
217
+ "f1": 0.703254,
218
+ "f1_weighted": 0.718177
219
+ },
220
+ {
221
+ "accuracy": 0.737334,
222
+ "f1": 0.723067,
223
+ "f1_weighted": 0.73131
224
+ },
225
+ {
226
+ "accuracy": 0.714707,
227
+ "f1": 0.700795,
228
+ "f1_weighted": 0.701288
229
+ },
230
+ {
231
+ "accuracy": 0.725037,
232
+ "f1": 0.70656,
233
+ "f1_weighted": 0.720189
234
+ },
235
+ {
236
+ "accuracy": 0.697,
237
+ "f1": 0.696585,
238
+ "f1_weighted": 0.694791
239
+ },
240
+ {
241
+ "accuracy": 0.740777,
242
+ "f1": 0.726883,
243
+ "f1_weighted": 0.737555
244
+ },
245
+ {
246
+ "accuracy": 0.736842,
247
+ "f1": 0.718202,
248
+ "f1_weighted": 0.728979
249
+ }
250
+ ],
251
+ "main_score": 0.728037,
252
+ "hf_subset": "de",
253
+ "languages": [
254
+ "deu-Latn"
255
+ ]
256
  }
257
  ]
258
  },
259
+ "evaluation_time": 34.10709095001221,
260
+ "kg_co2_emissions": null
261
  }
results/BAAI__bge-m3/5617a9f61b028005a4858fdac845db406aefb181/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "BAAI/bge-m3", "revision": "5617a9f61b028005a4858fdac845db406aefb181", "release_date": "2024-02-05", "languages": ["eng_Latn"], "n_parameters": null, "memory_usage": null, "max_tokens": null, "embed_dim": null, "license": null, "open_weights": true, "public_training_data": null, "public_training_code": null, "framework": [], "reference": null, "similarity_fn_name": null, "use_instructions": null, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "BGEM3Wrapper"}
 
1
+ {"name": "BAAI/bge-m3", "revision": "5617a9f61b028005a4858fdac845db406aefb181", "release_date": "2024-06-28", "languages": ["afr_Latn", "amh_Ethi", "azj_Latn", "ast_Latn", "azj_Latn", "ben_Beng", "bul_Cyrl", "bel_Cyrl", "cat_Latn", "ceb_Latn", "ckb_Arab", "dan_Latn", "deu_Latn", "ell_Grek", "eng_Latn", "est_Latn", "fin_Latn", "fra_Latn", "glg_Latn", "guj_Gujr", "heb_Hebr", "hin_Deva", "ita_Latn", "jpn_Jpan", "kor_Hang", "rus_Cyrl", "tha_Thai", "ukr_Cyrl", "zho_Hans"], "n_parameters": 568000000, "max_tokens": 8194.0, "embed_dim": 4096, "license": "mit", "open_weights": true, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/BAAI/bge-m3", "similarity_fn_name": "cosine", "use_instructions": false, "training_datasets": {"T2Retrieval": ["train"], "DuReader": ["train"], "MMarcoReranking": ["train"], "CMedQAv2-reranking": ["train"], "HotpotQA": ["train"], "NQ": ["train"], "MSMARCO": ["train"], "MrTidyRetrieval": ["train"], "MIRACLRetrieval": ["train"], "CodeSearchNet": ["train"]}, "adapted_from": null, "superseded_by": null, "loader": "sentence_transformers_loader"}
results/Cohere__Cohere-embed-multilingual-light-v3.0/1/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "Cohere/Cohere-embed-multilingual-light-v3.0", "revision": "1", "release_date": "2023-11-02", "languages": ["afr-Latn", "amh-Ethi", "ara-Arab", "asm-Beng", "aze-Latn", "bel-Cyrl", "bul-Cyrl", "ben-Beng", "bod-Tibt", "bos-Latn", "cat-Latn", "ceb-Latn", "cos-Latn", "ces-Latn", "cym-Latn", "dan-Latn", "deu-Latn", "ell-Grek", "eng-Latn", "epo-Latn", "spa-Latn", "est-Latn", "eus-Latn", "fas-Arab", "fin-Latn", "fra-Latn", "fry-Latn", "gle-Latn", "gla-Latn", "glg-Latn", "guj-Gujr", "hau-Latn", "haw-Latn", "heb-Hebr", "hin-Deva", "hmn-Latn", "hrv-Latn", "hat-Latn", "hun-Latn", "hye-Armn", "ind-Latn", "ibo-Latn", "isl-Latn", "ita-Latn", "jpn-Jpan", "jav-Latn", "kat-Geor", "kaz-Cyrl", "khm-Khmr", "kan-Knda", "kor-Kore", "kur-Arab", "kir-Cyrl", "lat-Latn", "ltz-Latn", "lao-Laoo", "lit-Latn", "lav-Latn", "mlg-Latn", "mri-Latn", "mkd-Cyrl", "mal-Mlym", "mon-Cyrl", "mar-Deva", "msa-Latn", "mlt-Latn", "mya-Mymr", "nep-Deva", "nld-Latn", "nor-Latn", "nya-Latn", "ori-Orya", "pan-Guru", "pol-Latn", "por-Latn", "ron-Latn", "rus-Cyrl", "kin-Latn", "sin-Sinh", "slk-Latn", "slv-Latn", "smo-Latn", "sna-Latn", "som-Latn", "sqi-Latn", "srp-Cyrl", "sot-Latn", "sun-Latn", "swe-Latn", "swa-Latn", "tam-Taml", "tel-Telu", "tgk-Cyrl", "tha-Thai", "tuk-Latn", "tgl-Latn", "tur-Latn", "tat-Cyrl", "uig-Arab", "ukr-Cyrl", "urd-Arab", "uzb-Latn", "vie-Latn", "wol-Latn", "xho-Latn", "yid-Hebr", "yor-Latn", "zho-Hans", "zul-Latn"], "n_parameters": null, "memory_usage": null, "max_tokens": 512.0, "embed_dim": 384, "license": null, "open_weights": false, "public_training_data": null, "public_training_code": null, "framework": ["API"], "reference": "https://cohere.com/blog/introducing-embed-v3", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "CohereTextEmbeddingModel"}
 
1
+ {"name": "Cohere/Cohere-embed-multilingual-light-v3.0", "revision": "1", "release_date": "2023-11-02", "languages": ["afr-Latn", "amh-Ethi", "ara-Arab", "asm-Beng", "aze-Latn", "bel-Cyrl", "bul-Cyrl", "ben-Beng", "bod-Tibt", "bos-Latn", "cat-Latn", "ceb-Latn", "cos-Latn", "ces-Latn", "cym-Latn", "dan-Latn", "deu-Latn", "ell-Grek", "eng-Latn", "epo-Latn", "spa-Latn", "est-Latn", "eus-Latn", "fas-Arab", "fin-Latn", "fra-Latn", "fry-Latn", "gle-Latn", "gla-Latn", "glg-Latn", "guj-Gujr", "hau-Latn", "haw-Latn", "heb-Hebr", "hin-Deva", "hmn-Latn", "hrv-Latn", "hat-Latn", "hun-Latn", "hye-Armn", "ind-Latn", "ibo-Latn", "isl-Latn", "ita-Latn", "jpn-Jpan", "jav-Latn", "kat-Geor", "kaz-Cyrl", "khm-Khmr", "kan-Knda", "kor-Kore", "kur-Arab", "kir-Cyrl", "lat-Latn", "ltz-Latn", "lao-Laoo", "lit-Latn", "lav-Latn", "mlg-Latn", "mri-Latn", "mkd-Cyrl", "mal-Mlym", "mon-Cyrl", "mar-Deva", "msa-Latn", "mlt-Latn", "mya-Mymr", "nep-Deva", "nld-Latn", "nor-Latn", "nya-Latn", "ori-Orya", "pan-Guru", "pol-Latn", "por-Latn", "ron-Latn", "rus-Cyrl", "kin-Latn", "sin-Sinh", "slk-Latn", "slv-Latn", "smo-Latn", "sna-Latn", "som-Latn", "sqi-Latn", "srp-Cyrl", "sot-Latn", "sun-Latn", "swe-Latn", "swa-Latn", "tam-Taml", "tel-Telu", "tgk-Cyrl", "tha-Thai", "tuk-Latn", "tgl-Latn", "tur-Latn", "tat-Cyrl", "uig-Arab", "ukr-Cyrl", "urd-Arab", "uzb-Latn", "vie-Latn", "wol-Latn", "xho-Latn", "yid-Hebr", "yor-Latn", "zho-Hans", "zul-Latn"], "n_parameters": null, "max_tokens": 512.0, "embed_dim": 384, "license": null, "open_weights": false, "public_training_code": null, "framework": ["API"], "reference": "https://cohere.com/blog/introducing-embed-v3", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "CohereTextEmbeddingModel"}
results/Cohere__Cohere-embed-multilingual-v3.0/1/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "Cohere/Cohere-embed-multilingual-v3.0", "revision": "1", "release_date": "2023-11-02", "languages": ["afr-Latn", "amh-Ethi", "ara-Arab", "asm-Beng", "aze-Latn", "bel-Cyrl", "bul-Cyrl", "ben-Beng", "bod-Tibt", "bos-Latn", "cat-Latn", "ceb-Latn", "cos-Latn", "ces-Latn", "cym-Latn", "dan-Latn", "deu-Latn", "ell-Grek", "eng-Latn", "epo-Latn", "spa-Latn", "est-Latn", "eus-Latn", "fas-Arab", "fin-Latn", "fra-Latn", "fry-Latn", "gle-Latn", "gla-Latn", "glg-Latn", "guj-Gujr", "hau-Latn", "haw-Latn", "heb-Hebr", "hin-Deva", "hmn-Latn", "hrv-Latn", "hat-Latn", "hun-Latn", "hye-Armn", "ind-Latn", "ibo-Latn", "isl-Latn", "ita-Latn", "jpn-Jpan", "jav-Latn", "kat-Geor", "kaz-Cyrl", "khm-Khmr", "kan-Knda", "kor-Kore", "kur-Arab", "kir-Cyrl", "lat-Latn", "ltz-Latn", "lao-Laoo", "lit-Latn", "lav-Latn", "mlg-Latn", "mri-Latn", "mkd-Cyrl", "mal-Mlym", "mon-Cyrl", "mar-Deva", "msa-Latn", "mlt-Latn", "mya-Mymr", "nep-Deva", "nld-Latn", "nor-Latn", "nya-Latn", "ori-Orya", "pan-Guru", "pol-Latn", "por-Latn", "ron-Latn", "rus-Cyrl", "kin-Latn", "sin-Sinh", "slk-Latn", "slv-Latn", "smo-Latn", "sna-Latn", "som-Latn", "sqi-Latn", "srp-Cyrl", "sot-Latn", "sun-Latn", "swe-Latn", "swa-Latn", "tam-Taml", "tel-Telu", "tgk-Cyrl", "tha-Thai", "tuk-Latn", "tgl-Latn", "tur-Latn", "tat-Cyrl", "uig-Arab", "ukr-Cyrl", "urd-Arab", "uzb-Latn", "vie-Latn", "wol-Latn", "xho-Latn", "yid-Hebr", "yor-Latn", "zho-Hans", "zul-Latn"], "n_parameters": null, "memory_usage": null, "max_tokens": null, "embed_dim": 512, "license": null, "open_weights": false, "public_training_data": null, "public_training_code": null, "framework": ["API"], "reference": "https://cohere.com/blog/introducing-embed-v3", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "CohereTextEmbeddingModel"}
 
1
+ {"name": "Cohere/Cohere-embed-multilingual-v3.0", "revision": "1", "release_date": "2023-11-02", "languages": ["afr-Latn", "amh-Ethi", "ara-Arab", "asm-Beng", "aze-Latn", "bel-Cyrl", "bul-Cyrl", "ben-Beng", "bod-Tibt", "bos-Latn", "cat-Latn", "ceb-Latn", "cos-Latn", "ces-Latn", "cym-Latn", "dan-Latn", "deu-Latn", "ell-Grek", "eng-Latn", "epo-Latn", "spa-Latn", "est-Latn", "eus-Latn", "fas-Arab", "fin-Latn", "fra-Latn", "fry-Latn", "gle-Latn", "gla-Latn", "glg-Latn", "guj-Gujr", "hau-Latn", "haw-Latn", "heb-Hebr", "hin-Deva", "hmn-Latn", "hrv-Latn", "hat-Latn", "hun-Latn", "hye-Armn", "ind-Latn", "ibo-Latn", "isl-Latn", "ita-Latn", "jpn-Jpan", "jav-Latn", "kat-Geor", "kaz-Cyrl", "khm-Khmr", "kan-Knda", "kor-Kore", "kur-Arab", "kir-Cyrl", "lat-Latn", "ltz-Latn", "lao-Laoo", "lit-Latn", "lav-Latn", "mlg-Latn", "mri-Latn", "mkd-Cyrl", "mal-Mlym", "mon-Cyrl", "mar-Deva", "msa-Latn", "mlt-Latn", "mya-Mymr", "nep-Deva", "nld-Latn", "nor-Latn", "nya-Latn", "ori-Orya", "pan-Guru", "pol-Latn", "por-Latn", "ron-Latn", "rus-Cyrl", "kin-Latn", "sin-Sinh", "slk-Latn", "slv-Latn", "smo-Latn", "sna-Latn", "som-Latn", "sqi-Latn", "srp-Cyrl", "sot-Latn", "sun-Latn", "swe-Latn", "swa-Latn", "tam-Taml", "tel-Telu", "tgk-Cyrl", "tha-Thai", "tuk-Latn", "tgl-Latn", "tur-Latn", "tat-Cyrl", "uig-Arab", "ukr-Cyrl", "urd-Arab", "uzb-Latn", "vie-Latn", "wol-Latn", "xho-Latn", "yid-Hebr", "yor-Latn", "zho-Hans", "zul-Latn"], "n_parameters": null, "max_tokens": null, "embed_dim": 512, "license": null, "open_weights": false, "public_training_code": null, "framework": ["API"], "reference": "https://cohere.com/blog/introducing-embed-v3", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "CohereTextEmbeddingModel"}
results/Snowflake__snowflake-arctic-embed-l-v2.0/edc2df7b6c25794b340229ca082e7c78782e6374/GermanDPR.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "5129d02422a66be600ac89cd3e8531b4f97d347d",
3
+ "task_name": "GermanDPR",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.67512,
9
+ "ndcg_at_3": 0.80062,
10
+ "ndcg_at_5": 0.82704,
11
+ "ndcg_at_10": 0.83674,
12
+ "ndcg_at_20": 0.83976,
13
+ "ndcg_at_100": 0.84097,
14
+ "ndcg_at_1000": 0.84111,
15
+ "map_at_1": 0.67512,
16
+ "map_at_3": 0.77041,
17
+ "map_at_5": 0.78524,
18
+ "map_at_10": 0.78937,
19
+ "map_at_20": 0.79024,
20
+ "map_at_100": 0.79038,
21
+ "map_at_1000": 0.79039,
22
+ "recall_at_1": 0.67512,
23
+ "recall_at_3": 0.8878,
24
+ "recall_at_5": 0.95122,
25
+ "recall_at_10": 0.98049,
26
+ "recall_at_20": 0.9922,
27
+ "recall_at_100": 0.99902,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.67512,
30
+ "precision_at_3": 0.29593,
31
+ "precision_at_5": 0.19024,
32
+ "precision_at_10": 0.09805,
33
+ "precision_at_20": 0.04961,
34
+ "precision_at_100": 0.00999,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.674146,
37
+ "mrr_at_3": 0.769919,
38
+ "mrr_at_5": 0.784748,
39
+ "mrr_at_10": 0.788886,
40
+ "mrr_at_20": 0.78975,
41
+ "mrr_at_100": 0.789896,
42
+ "mrr_at_1000": 0.789905,
43
+ "nauc_ndcg_at_1_max": 0.116589,
44
+ "nauc_ndcg_at_1_std": -0.132546,
45
+ "nauc_ndcg_at_1_diff1": 0.526583,
46
+ "nauc_ndcg_at_3_max": 0.175733,
47
+ "nauc_ndcg_at_3_std": -0.124489,
48
+ "nauc_ndcg_at_3_diff1": 0.485141,
49
+ "nauc_ndcg_at_5_max": 0.166985,
50
+ "nauc_ndcg_at_5_std": -0.126915,
51
+ "nauc_ndcg_at_5_diff1": 0.507238,
52
+ "nauc_ndcg_at_10_max": 0.174544,
53
+ "nauc_ndcg_at_10_std": -0.105128,
54
+ "nauc_ndcg_at_10_diff1": 0.50522,
55
+ "nauc_ndcg_at_20_max": 0.162368,
56
+ "nauc_ndcg_at_20_std": -0.114208,
57
+ "nauc_ndcg_at_20_diff1": 0.502748,
58
+ "nauc_ndcg_at_100_max": 0.157911,
59
+ "nauc_ndcg_at_100_std": -0.119919,
60
+ "nauc_ndcg_at_100_diff1": 0.504159,
61
+ "nauc_ndcg_at_1000_max": 0.157031,
62
+ "nauc_ndcg_at_1000_std": -0.121089,
63
+ "nauc_ndcg_at_1000_diff1": 0.505275,
64
+ "nauc_map_at_1_max": 0.116589,
65
+ "nauc_map_at_1_std": -0.132546,
66
+ "nauc_map_at_1_diff1": 0.526583,
67
+ "nauc_map_at_3_max": 0.157563,
68
+ "nauc_map_at_3_std": -0.127467,
69
+ "nauc_map_at_3_diff1": 0.496652,
70
+ "nauc_map_at_5_max": 0.151622,
71
+ "nauc_map_at_5_std": -0.128939,
72
+ "nauc_map_at_5_diff1": 0.507892,
73
+ "nauc_map_at_10_max": 0.154139,
74
+ "nauc_map_at_10_std": -0.121437,
75
+ "nauc_map_at_10_diff1": 0.507155,
76
+ "nauc_map_at_20_max": 0.151407,
77
+ "nauc_map_at_20_std": -0.123781,
78
+ "nauc_map_at_20_diff1": 0.506651,
79
+ "nauc_map_at_100_max": 0.151002,
80
+ "nauc_map_at_100_std": -0.124382,
81
+ "nauc_map_at_100_diff1": 0.506897,
82
+ "nauc_map_at_1000_max": 0.150959,
83
+ "nauc_map_at_1000_std": -0.124439,
84
+ "nauc_map_at_1000_diff1": 0.506951,
85
+ "nauc_recall_at_1_max": 0.116589,
86
+ "nauc_recall_at_1_std": -0.132546,
87
+ "nauc_recall_at_1_diff1": 0.526583,
88
+ "nauc_recall_at_3_max": 0.27038,
89
+ "nauc_recall_at_3_std": -0.108572,
90
+ "nauc_recall_at_3_diff1": 0.426163,
91
+ "nauc_recall_at_5_max": 0.333143,
92
+ "nauc_recall_at_5_std": -0.104512,
93
+ "nauc_recall_at_5_diff1": 0.513295,
94
+ "nauc_recall_at_10_max": 0.732287,
95
+ "nauc_recall_at_10_std": 0.395433,
96
+ "nauc_recall_at_10_diff1": 0.481916,
97
+ "nauc_recall_at_20_max": 0.718749,
98
+ "nauc_recall_at_20_std": 0.564973,
99
+ "nauc_recall_at_20_diff1": 0.270563,
100
+ "nauc_recall_at_100_max": 1.0,
101
+ "nauc_recall_at_100_std": 1.0,
102
+ "nauc_recall_at_100_diff1": -0.563948,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.116589,
107
+ "nauc_precision_at_1_std": -0.132546,
108
+ "nauc_precision_at_1_diff1": 0.526583,
109
+ "nauc_precision_at_3_max": 0.27038,
110
+ "nauc_precision_at_3_std": -0.108572,
111
+ "nauc_precision_at_3_diff1": 0.426163,
112
+ "nauc_precision_at_5_max": 0.333143,
113
+ "nauc_precision_at_5_std": -0.104512,
114
+ "nauc_precision_at_5_diff1": 0.513295,
115
+ "nauc_precision_at_10_max": 0.732287,
116
+ "nauc_precision_at_10_std": 0.395433,
117
+ "nauc_precision_at_10_diff1": 0.481916,
118
+ "nauc_precision_at_20_max": 0.718749,
119
+ "nauc_precision_at_20_std": 0.564973,
120
+ "nauc_precision_at_20_diff1": 0.270563,
121
+ "nauc_precision_at_100_max": 1.0,
122
+ "nauc_precision_at_100_std": 1.0,
123
+ "nauc_precision_at_100_diff1": -0.563948,
124
+ "nauc_precision_at_1000_max": NaN,
125
+ "nauc_precision_at_1000_std": NaN,
126
+ "nauc_precision_at_1000_diff1": NaN,
127
+ "nauc_mrr_at_1_max": 0.112249,
128
+ "nauc_mrr_at_1_std": -0.138482,
129
+ "nauc_mrr_at_1_diff1": 0.52911,
130
+ "nauc_mrr_at_3_max": 0.154564,
131
+ "nauc_mrr_at_3_std": -0.1315,
132
+ "nauc_mrr_at_3_diff1": 0.498405,
133
+ "nauc_mrr_at_5_max": 0.148401,
134
+ "nauc_mrr_at_5_std": -0.13328,
135
+ "nauc_mrr_at_5_diff1": 0.509765,
136
+ "nauc_mrr_at_10_max": 0.150842,
137
+ "nauc_mrr_at_10_std": -0.125888,
138
+ "nauc_mrr_at_10_diff1": 0.509071,
139
+ "nauc_mrr_at_20_max": 0.148098,
140
+ "nauc_mrr_at_20_std": -0.128249,
141
+ "nauc_mrr_at_20_diff1": 0.508576,
142
+ "nauc_mrr_at_100_max": 0.147691,
143
+ "nauc_mrr_at_100_std": -0.128853,
144
+ "nauc_mrr_at_100_diff1": 0.508823,
145
+ "nauc_mrr_at_1000_max": 0.147648,
146
+ "nauc_mrr_at_1000_std": -0.128909,
147
+ "nauc_mrr_at_1000_diff1": 0.508877,
148
+ "main_score": 0.83674,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 203.90697646141052,
157
+ "kg_co2_emissions": null
158
+ }
results/Snowflake__snowflake-arctic-embed-l-v2.0/edc2df7b6c25794b340229ca082e7c78782e6374/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "Snowflake/snowflake-arctic-embed-l-v2.0", "revision": "edc2df7b6c25794b340229ca082e7c78782e6374", "release_date": "2024-12-04", "languages": ["afr_Latn", "ara_Arab", "aze_Latn", "bel_Cyrl", "bul_Cyrl", "ben_Beng", "cat_Latn", "ceb_Latn", "ces_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "ell_Grek", "eng_Latn", "spa_Latn", "est_Latn", "eus_Latn", "fas_Arab", "fin_Latn", "fra_Latn", "glg_Latn", "guj_Gujr", "heb_Hebr", "hin_Deva", "hrv_Latn", "hat_Latn", "hun_Latn", "hye_Armn", "ind_Latn", "isl_Latn", "ita_Latn", "jpn_Jpan", "jav_Latn", "kat_Geor", "kaz_Cyrl", "khm_Khmr", "kan_Knda", "kor_Hang", "kir_Cyrl", "lao_Laoo", "lit_Latn", "lav_Latn", "mkd_Cyrl", "mal_Mlym", "mon_Cyrl", "mar_Deva", "msa_Latn", "mya_Mymr", "nep_Deva", "nld_Latn", "pan_Guru", "pol_Latn", "por_Latn", "que_Latn", "ron_Latn", "rus_Cyrl", "sin_Sinh", "slk_Latn", "slv_Latn", "som_Latn", "sqi_Latn", "srp_Cyrl", "swe_Latn", "swa_Latn", "tam_Taml", "tel_Telu", "tha_Thai", "tgl_Latn", "tur_Latn", "ukr_Cyrl", "urd_Arab", "vie_Latn", "yor_Latn", "zho_Hans"], "n_parameters": 568000000, "memory_usage": null, "max_tokens": 8192.0, "embed_dim": 1024, "license": "apache-2.0", "open_weights": true, "public_training_data": null, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/Snowflake/snowflake-arctic-embed-l-v2.0", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": "BAAI/bge-m3-retromae", "superseded_by": null, "loader": "sentence_transformers_loader"}
 
1
+ {"name": "Snowflake/snowflake-arctic-embed-l-v2.0", "revision": "edc2df7b6c25794b340229ca082e7c78782e6374", "release_date": "2024-12-04", "languages": ["afr_Latn", "ara_Arab", "aze_Latn", "bel_Cyrl", "bul_Cyrl", "ben_Beng", "cat_Latn", "ceb_Latn", "ces_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "ell_Grek", "eng_Latn", "spa_Latn", "est_Latn", "eus_Latn", "fas_Arab", "fin_Latn", "fra_Latn", "glg_Latn", "guj_Gujr", "heb_Hebr", "hin_Deva", "hrv_Latn", "hat_Latn", "hun_Latn", "hye_Armn", "ind_Latn", "isl_Latn", "ita_Latn", "jpn_Jpan", "jav_Latn", "kat_Geor", "kaz_Cyrl", "khm_Khmr", "kan_Knda", "kor_Hang", "kir_Cyrl", "lao_Laoo", "lit_Latn", "lav_Latn", "mkd_Cyrl", "mal_Mlym", "mon_Cyrl", "mar_Deva", "msa_Latn", "mya_Mymr", "nep_Deva", "nld_Latn", "pan_Guru", "pol_Latn", "por_Latn", "que_Latn", "ron_Latn", "rus_Cyrl", "sin_Sinh", "slk_Latn", "slv_Latn", "som_Latn", "sqi_Latn", "srp_Cyrl", "swe_Latn", "swa_Latn", "tam_Taml", "tel_Telu", "tha_Thai", "tgl_Latn", "tur_Latn", "ukr_Cyrl", "urd_Arab", "vie_Latn", "yor_Latn", "zho_Hans"], "n_parameters": 568000000, "max_tokens": 8192.0, "embed_dim": 1024, "license": "apache-2.0", "open_weights": true, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/Snowflake/snowflake-arctic-embed-l-v2.0", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": {"NQ": ["test"], "NQHardNegatives": ["test"], "HotPotQA": ["test"], "HotPotQAHardNegatives": ["test"], "HotPotQA-PL": ["test"], "FEVER": ["test"], "FEVERHardNegatives": ["test"]}, "adapted_from": "BAAI/bge-m3-retromae", "superseded_by": null, "loader": "sentence_transformers_loader"}
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/AmazonCounterfactualClassification.json ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
3
+ "task_name": "AmazonCounterfactualClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.705794,
9
+ "f1": 0.68743,
10
+ "f1_weighted": 0.716737,
11
+ "ap": 0.814942,
12
+ "ap_weighted": 0.814942,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.738197,
16
+ "f1": 0.722147,
17
+ "f1_weighted": 0.748515,
18
+ "ap": 0.840371,
19
+ "ap_weighted": 0.840371
20
+ },
21
+ {
22
+ "accuracy": 0.736052,
23
+ "f1": 0.710274,
24
+ "f1_weighted": 0.744397,
25
+ "ap": 0.81851,
26
+ "ap_weighted": 0.81851
27
+ },
28
+ {
29
+ "accuracy": 0.759657,
30
+ "f1": 0.731876,
31
+ "f1_weighted": 0.765954,
32
+ "ap": 0.828119,
33
+ "ap_weighted": 0.828119
34
+ },
35
+ {
36
+ "accuracy": 0.654506,
37
+ "f1": 0.642616,
38
+ "f1_weighted": 0.668355,
39
+ "ap": 0.796669,
40
+ "ap_weighted": 0.796669
41
+ },
42
+ {
43
+ "accuracy": 0.718884,
44
+ "f1": 0.705028,
45
+ "f1_weighted": 0.730271,
46
+ "ap": 0.83386,
47
+ "ap_weighted": 0.83386
48
+ },
49
+ {
50
+ "accuracy": 0.688841,
51
+ "f1": 0.670125,
52
+ "f1_weighted": 0.70115,
53
+ "ap": 0.803576,
54
+ "ap_weighted": 0.803576
55
+ },
56
+ {
57
+ "accuracy": 0.708155,
58
+ "f1": 0.688159,
59
+ "f1_weighted": 0.719338,
60
+ "ap": 0.81283,
61
+ "ap_weighted": 0.81283
62
+ },
63
+ {
64
+ "accuracy": 0.639485,
65
+ "f1": 0.627965,
66
+ "f1_weighted": 0.653814,
67
+ "ap": 0.787802,
68
+ "ap_weighted": 0.787802
69
+ },
70
+ {
71
+ "accuracy": 0.751073,
72
+ "f1": 0.726376,
73
+ "f1_weighted": 0.758835,
74
+ "ap": 0.828942,
75
+ "ap_weighted": 0.828942
76
+ },
77
+ {
78
+ "accuracy": 0.66309,
79
+ "f1": 0.649733,
80
+ "f1_weighted": 0.676741,
81
+ "ap": 0.798744,
82
+ "ap_weighted": 0.798744
83
+ }
84
+ ],
85
+ "main_score": 0.705794,
86
+ "hf_subset": "de",
87
+ "languages": [
88
+ "deu-Latn"
89
+ ]
90
+ }
91
+ ],
92
+ "test": [
93
+ {
94
+ "accuracy": 0.698501,
95
+ "f1": 0.68184,
96
+ "f1_weighted": 0.709785,
97
+ "ap": 0.812608,
98
+ "ap_weighted": 0.812608,
99
+ "scores_per_experiment": [
100
+ {
101
+ "accuracy": 0.702355,
102
+ "f1": 0.686188,
103
+ "f1_weighted": 0.7141,
104
+ "ap": 0.815494,
105
+ "ap_weighted": 0.815494
106
+ },
107
+ {
108
+ "accuracy": 0.711991,
109
+ "f1": 0.688399,
110
+ "f1_weighted": 0.721998,
111
+ "ap": 0.806709,
112
+ "ap_weighted": 0.806709
113
+ },
114
+ {
115
+ "accuracy": 0.760171,
116
+ "f1": 0.734972,
117
+ "f1_weighted": 0.766996,
118
+ "ap": 0.831388,
119
+ "ap_weighted": 0.831388
120
+ },
121
+ {
122
+ "accuracy": 0.692719,
123
+ "f1": 0.681273,
124
+ "f1_weighted": 0.704942,
125
+ "ap": 0.821292,
126
+ "ap_weighted": 0.821292
127
+ },
128
+ {
129
+ "accuracy": 0.705567,
130
+ "f1": 0.694079,
131
+ "f1_weighted": 0.71731,
132
+ "ap": 0.829876,
133
+ "ap_weighted": 0.829876
134
+ },
135
+ {
136
+ "accuracy": 0.718415,
137
+ "f1": 0.701393,
138
+ "f1_weighted": 0.729331,
139
+ "ap": 0.823653,
140
+ "ap_weighted": 0.823653
141
+ },
142
+ {
143
+ "accuracy": 0.665953,
144
+ "f1": 0.647078,
145
+ "f1_weighted": 0.679061,
146
+ "ap": 0.787511,
147
+ "ap_weighted": 0.787511
148
+ },
149
+ {
150
+ "accuracy": 0.642398,
151
+ "f1": 0.632931,
152
+ "f1_weighted": 0.656032,
153
+ "ap": 0.792926,
154
+ "ap_weighted": 0.792926
155
+ },
156
+ {
157
+ "accuracy": 0.731263,
158
+ "f1": 0.708501,
159
+ "f1_weighted": 0.740421,
160
+ "ap": 0.819467,
161
+ "ap_weighted": 0.819467
162
+ },
163
+ {
164
+ "accuracy": 0.654176,
165
+ "f1": 0.643585,
166
+ "f1_weighted": 0.66766,
167
+ "ap": 0.797767,
168
+ "ap_weighted": 0.797767
169
+ }
170
+ ],
171
+ "main_score": 0.698501,
172
+ "hf_subset": "de",
173
+ "languages": [
174
+ "deu-Latn"
175
+ ]
176
+ }
177
+ ]
178
+ },
179
+ "evaluation_time": 85.73426079750061,
180
+ "kg_co2_emissions": null
181
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/AmazonReviewsClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1399c76144fd37290681b995c656ef9b2e06e26d",
3
+ "task_name": "AmazonReviewsClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.40684,
9
+ "f1": 0.385448,
10
+ "f1_weighted": 0.385448,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.444,
14
+ "f1": 0.433799,
15
+ "f1_weighted": 0.433799
16
+ },
17
+ {
18
+ "accuracy": 0.3936,
19
+ "f1": 0.373416,
20
+ "f1_weighted": 0.373416
21
+ },
22
+ {
23
+ "accuracy": 0.3854,
24
+ "f1": 0.374231,
25
+ "f1_weighted": 0.374231
26
+ },
27
+ {
28
+ "accuracy": 0.4334,
29
+ "f1": 0.400241,
30
+ "f1_weighted": 0.400241
31
+ },
32
+ {
33
+ "accuracy": 0.4202,
34
+ "f1": 0.38621,
35
+ "f1_weighted": 0.38621
36
+ },
37
+ {
38
+ "accuracy": 0.4074,
39
+ "f1": 0.383979,
40
+ "f1_weighted": 0.383979
41
+ },
42
+ {
43
+ "accuracy": 0.3426,
44
+ "f1": 0.330814,
45
+ "f1_weighted": 0.330814
46
+ },
47
+ {
48
+ "accuracy": 0.4326,
49
+ "f1": 0.402513,
50
+ "f1_weighted": 0.402513
51
+ },
52
+ {
53
+ "accuracy": 0.3986,
54
+ "f1": 0.375458,
55
+ "f1_weighted": 0.375458
56
+ },
57
+ {
58
+ "accuracy": 0.4106,
59
+ "f1": 0.393819,
60
+ "f1_weighted": 0.393819
61
+ }
62
+ ],
63
+ "main_score": 0.40684,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.40142,
73
+ "f1": 0.380438,
74
+ "f1_weighted": 0.380438,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.4248,
78
+ "f1": 0.415297,
79
+ "f1_weighted": 0.415297
80
+ },
81
+ {
82
+ "accuracy": 0.411,
83
+ "f1": 0.392726,
84
+ "f1_weighted": 0.392726
85
+ },
86
+ {
87
+ "accuracy": 0.3828,
88
+ "f1": 0.371937,
89
+ "f1_weighted": 0.371937
90
+ },
91
+ {
92
+ "accuracy": 0.4214,
93
+ "f1": 0.38879,
94
+ "f1_weighted": 0.38879
95
+ },
96
+ {
97
+ "accuracy": 0.4134,
98
+ "f1": 0.377413,
99
+ "f1_weighted": 0.377413
100
+ },
101
+ {
102
+ "accuracy": 0.39,
103
+ "f1": 0.364907,
104
+ "f1_weighted": 0.364907
105
+ },
106
+ {
107
+ "accuracy": 0.3464,
108
+ "f1": 0.332527,
109
+ "f1_weighted": 0.332527
110
+ },
111
+ {
112
+ "accuracy": 0.4312,
113
+ "f1": 0.404496,
114
+ "f1_weighted": 0.404496
115
+ },
116
+ {
117
+ "accuracy": 0.3938,
118
+ "f1": 0.369978,
119
+ "f1_weighted": 0.369978
120
+ },
121
+ {
122
+ "accuracy": 0.3994,
123
+ "f1": 0.386307,
124
+ "f1_weighted": 0.386307
125
+ }
126
+ ],
127
+ "main_score": 0.40142,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 39.509817361831665,
136
+ "kg_co2_emissions": null
137
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/FalseFriendsGermanEnglish.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "15d6c030d3336cbb09de97b2cefc46db93262d40",
3
+ "task_name": "FalseFriendsGermanEnglish",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "similarity_accuracy": 0.50853,
9
+ "similarity_accuracy_threshold": 0.961091,
10
+ "similarity_f1": 0.666958,
11
+ "similarity_f1_threshold": 0.737133,
12
+ "similarity_precision": 0.500328,
13
+ "similarity_recall": 1.0,
14
+ "similarity_ap": 0.496637,
15
+ "cosine_accuracy": 0.50853,
16
+ "cosine_accuracy_threshold": 0.96109,
17
+ "cosine_f1": 0.666958,
18
+ "cosine_f1_threshold": 0.737133,
19
+ "cosine_precision": 0.500328,
20
+ "cosine_recall": 1.0,
21
+ "cosine_ap": 0.496637,
22
+ "manhattan_accuracy": 0.50853,
23
+ "manhattan_accuracy_threshold": 3.618259,
24
+ "manhattan_f1": 0.666959,
25
+ "manhattan_f1_threshold": 11.207991,
26
+ "manhattan_precision": 0.500658,
27
+ "manhattan_recall": 0.998688,
28
+ "manhattan_ap": 0.496999,
29
+ "euclidean_accuracy": 0.50853,
30
+ "euclidean_accuracy_threshold": 0.278961,
31
+ "euclidean_f1": 0.666958,
32
+ "euclidean_f1_threshold": 0.725073,
33
+ "euclidean_precision": 0.500328,
34
+ "euclidean_recall": 1.0,
35
+ "euclidean_ap": 0.496637,
36
+ "dot_accuracy": 0.50853,
37
+ "dot_accuracy_threshold": 0.96109,
38
+ "dot_f1": 0.666958,
39
+ "dot_f1_threshold": 0.737133,
40
+ "dot_precision": 0.500328,
41
+ "dot_recall": 1.0,
42
+ "dot_ap": 0.496637,
43
+ "max_accuracy": 0.50853,
44
+ "max_f1": 0.666959,
45
+ "max_precision": 0.500658,
46
+ "max_recall": 1.0,
47
+ "max_ap": 0.496999,
48
+ "main_score": 0.496999,
49
+ "hf_subset": "default",
50
+ "languages": [
51
+ "deu-Latn"
52
+ ]
53
+ }
54
+ ]
55
+ },
56
+ "evaluation_time": 1.0718636512756348,
57
+ "kg_co2_emissions": null
58
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/GermanDPR.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "5129d02422a66be600ac89cd3e8531b4f97d347d",
3
+ "task_name": "GermanDPR",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.5922,
9
+ "ndcg_at_3": 0.74547,
10
+ "ndcg_at_5": 0.77519,
11
+ "ndcg_at_10": 0.78938,
12
+ "ndcg_at_20": 0.79401,
13
+ "ndcg_at_100": 0.79528,
14
+ "ndcg_at_1000": 0.79581,
15
+ "map_at_1": 0.5922,
16
+ "map_at_3": 0.70715,
17
+ "map_at_5": 0.72364,
18
+ "map_at_10": 0.72968,
19
+ "map_at_20": 0.73105,
20
+ "map_at_100": 0.73123,
21
+ "map_at_1000": 0.73125,
22
+ "recall_at_1": 0.5922,
23
+ "recall_at_3": 0.85659,
24
+ "recall_at_5": 0.92878,
25
+ "recall_at_10": 0.97171,
26
+ "recall_at_20": 0.98927,
27
+ "recall_at_100": 0.9961,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.5922,
30
+ "precision_at_3": 0.28553,
31
+ "precision_at_5": 0.18576,
32
+ "precision_at_10": 0.09717,
33
+ "precision_at_20": 0.04946,
34
+ "precision_at_100": 0.00996,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.592195,
37
+ "mrr_at_3": 0.707154,
38
+ "mrr_at_5": 0.723642,
39
+ "mrr_at_10": 0.729678,
40
+ "mrr_at_20": 0.731047,
41
+ "mrr_at_100": 0.731231,
42
+ "mrr_at_1000": 0.731255,
43
+ "nauc_ndcg_at_1_max": 0.203407,
44
+ "nauc_ndcg_at_1_std": -0.126544,
45
+ "nauc_ndcg_at_1_diff1": 0.521519,
46
+ "nauc_ndcg_at_3_max": 0.263472,
47
+ "nauc_ndcg_at_3_std": -0.129686,
48
+ "nauc_ndcg_at_3_diff1": 0.460559,
49
+ "nauc_ndcg_at_5_max": 0.266253,
50
+ "nauc_ndcg_at_5_std": -0.119333,
51
+ "nauc_ndcg_at_5_diff1": 0.461402,
52
+ "nauc_ndcg_at_10_max": 0.26549,
53
+ "nauc_ndcg_at_10_std": -0.098176,
54
+ "nauc_ndcg_at_10_diff1": 0.474746,
55
+ "nauc_ndcg_at_20_max": 0.25329,
56
+ "nauc_ndcg_at_20_std": -0.100426,
57
+ "nauc_ndcg_at_20_diff1": 0.475069,
58
+ "nauc_ndcg_at_100_max": 0.249986,
59
+ "nauc_ndcg_at_100_std": -0.104688,
60
+ "nauc_ndcg_at_100_diff1": 0.477726,
61
+ "nauc_ndcg_at_1000_max": 0.249041,
62
+ "nauc_ndcg_at_1000_std": -0.108078,
63
+ "nauc_ndcg_at_1000_diff1": 0.478287,
64
+ "nauc_map_at_1_max": 0.203407,
65
+ "nauc_map_at_1_std": -0.126544,
66
+ "nauc_map_at_1_diff1": 0.521519,
67
+ "nauc_map_at_3_max": 0.246318,
68
+ "nauc_map_at_3_std": -0.125108,
69
+ "nauc_map_at_3_diff1": 0.476715,
70
+ "nauc_map_at_5_max": 0.247093,
71
+ "nauc_map_at_5_std": -0.119102,
72
+ "nauc_map_at_5_diff1": 0.478153,
73
+ "nauc_map_at_10_max": 0.246361,
74
+ "nauc_map_at_10_std": -0.111469,
75
+ "nauc_map_at_10_diff1": 0.482702,
76
+ "nauc_map_at_20_max": 0.24351,
77
+ "nauc_map_at_20_std": -0.111945,
78
+ "nauc_map_at_20_diff1": 0.482763,
79
+ "nauc_map_at_100_max": 0.243109,
80
+ "nauc_map_at_100_std": -0.1125,
81
+ "nauc_map_at_100_diff1": 0.482991,
82
+ "nauc_map_at_1000_max": 0.243075,
83
+ "nauc_map_at_1000_std": -0.112617,
84
+ "nauc_map_at_1000_diff1": 0.48302,
85
+ "nauc_recall_at_1_max": 0.203407,
86
+ "nauc_recall_at_1_std": -0.126544,
87
+ "nauc_recall_at_1_diff1": 0.521519,
88
+ "nauc_recall_at_3_max": 0.348888,
89
+ "nauc_recall_at_3_std": -0.155436,
90
+ "nauc_recall_at_3_diff1": 0.381121,
91
+ "nauc_recall_at_5_max": 0.437975,
92
+ "nauc_recall_at_5_std": -0.125614,
93
+ "nauc_recall_at_5_diff1": 0.312717,
94
+ "nauc_recall_at_10_max": 0.682578,
95
+ "nauc_recall_at_10_std": 0.234242,
96
+ "nauc_recall_at_10_diff1": 0.350383,
97
+ "nauc_recall_at_20_max": 0.628913,
98
+ "nauc_recall_at_20_std": 0.624248,
99
+ "nauc_recall_at_20_diff1": 0.175382,
100
+ "nauc_recall_at_100_max": 0.546719,
101
+ "nauc_recall_at_100_std": 0.967338,
102
+ "nauc_recall_at_100_diff1": 0.320328,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.203407,
107
+ "nauc_precision_at_1_std": -0.126544,
108
+ "nauc_precision_at_1_diff1": 0.521519,
109
+ "nauc_precision_at_3_max": 0.348888,
110
+ "nauc_precision_at_3_std": -0.155436,
111
+ "nauc_precision_at_3_diff1": 0.381121,
112
+ "nauc_precision_at_5_max": 0.437975,
113
+ "nauc_precision_at_5_std": -0.125614,
114
+ "nauc_precision_at_5_diff1": 0.312717,
115
+ "nauc_precision_at_10_max": 0.682578,
116
+ "nauc_precision_at_10_std": 0.234242,
117
+ "nauc_precision_at_10_diff1": 0.350383,
118
+ "nauc_precision_at_20_max": 0.628913,
119
+ "nauc_precision_at_20_std": 0.624248,
120
+ "nauc_precision_at_20_diff1": 0.175382,
121
+ "nauc_precision_at_100_max": 0.546719,
122
+ "nauc_precision_at_100_std": 0.967338,
123
+ "nauc_precision_at_100_diff1": 0.320328,
124
+ "nauc_precision_at_1000_max": NaN,
125
+ "nauc_precision_at_1000_std": NaN,
126
+ "nauc_precision_at_1000_diff1": NaN,
127
+ "nauc_mrr_at_1_max": 0.203407,
128
+ "nauc_mrr_at_1_std": -0.126544,
129
+ "nauc_mrr_at_1_diff1": 0.521519,
130
+ "nauc_mrr_at_3_max": 0.246318,
131
+ "nauc_mrr_at_3_std": -0.125108,
132
+ "nauc_mrr_at_3_diff1": 0.476715,
133
+ "nauc_mrr_at_5_max": 0.247093,
134
+ "nauc_mrr_at_5_std": -0.119102,
135
+ "nauc_mrr_at_5_diff1": 0.478153,
136
+ "nauc_mrr_at_10_max": 0.246361,
137
+ "nauc_mrr_at_10_std": -0.111469,
138
+ "nauc_mrr_at_10_diff1": 0.482702,
139
+ "nauc_mrr_at_20_max": 0.24351,
140
+ "nauc_mrr_at_20_std": -0.111945,
141
+ "nauc_mrr_at_20_diff1": 0.482763,
142
+ "nauc_mrr_at_100_max": 0.243109,
143
+ "nauc_mrr_at_100_std": -0.1125,
144
+ "nauc_mrr_at_100_diff1": 0.482991,
145
+ "nauc_mrr_at_1000_max": 0.243075,
146
+ "nauc_mrr_at_1000_std": -0.112617,
147
+ "nauc_mrr_at_1000_diff1": 0.48302,
148
+ "main_score": 0.78938,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 13.16852593421936,
157
+ "kg_co2_emissions": null
158
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/GermanQuAD-Retrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f5c87ae5a2e7a5106606314eef45255f03151bb3",
3
+ "task_name": "GermanQuAD-Retrieval",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.87205,
9
+ "ndcg_at_3": 0.92325,
10
+ "ndcg_at_5": 0.92743,
11
+ "ndcg_at_10": 0.9314,
12
+ "ndcg_at_20": 0.93313,
13
+ "ndcg_at_100": 0.93528,
14
+ "ndcg_at_1000": 0.93559,
15
+ "map_at_1": 0.87205,
16
+ "map_at_3": 0.91137,
17
+ "map_at_5": 0.91373,
18
+ "map_at_10": 0.91537,
19
+ "map_at_20": 0.91585,
20
+ "map_at_100": 0.91617,
21
+ "map_at_1000": 0.91618,
22
+ "recall_at_1": 0.87205,
23
+ "recall_at_3": 0.95735,
24
+ "recall_at_5": 0.96733,
25
+ "recall_at_10": 0.97958,
26
+ "recall_at_20": 0.98639,
27
+ "recall_at_100": 0.99773,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.87205,
30
+ "precision_at_3": 0.31912,
31
+ "precision_at_5": 0.19347,
32
+ "precision_at_10": 0.09796,
33
+ "precision_at_20": 0.04932,
34
+ "precision_at_100": 0.00998,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.872051,
37
+ "mrr_at_3": 0.911373,
38
+ "mrr_at_5": 0.913733,
39
+ "mrr_at_10": 0.915374,
40
+ "mrr_at_20": 0.915852,
41
+ "mrr_at_100": 0.916166,
42
+ "mrr_at_1000": 0.916182,
43
+ "nauc_ndcg_at_1_max": 0.715384,
44
+ "nauc_ndcg_at_1_std": -0.153748,
45
+ "nauc_ndcg_at_1_diff1": 0.899327,
46
+ "nauc_ndcg_at_3_max": 0.778567,
47
+ "nauc_ndcg_at_3_std": -0.136698,
48
+ "nauc_ndcg_at_3_diff1": 0.887663,
49
+ "nauc_ndcg_at_5_max": 0.787372,
50
+ "nauc_ndcg_at_5_std": -0.130923,
51
+ "nauc_ndcg_at_5_diff1": 0.891482,
52
+ "nauc_ndcg_at_10_max": 0.778647,
53
+ "nauc_ndcg_at_10_std": -0.131639,
54
+ "nauc_ndcg_at_10_diff1": 0.893914,
55
+ "nauc_ndcg_at_20_max": 0.773984,
56
+ "nauc_ndcg_at_20_std": -0.134365,
57
+ "nauc_ndcg_at_20_diff1": 0.893319,
58
+ "nauc_ndcg_at_100_max": 0.766479,
59
+ "nauc_ndcg_at_100_std": -0.141911,
60
+ "nauc_ndcg_at_100_diff1": 0.893434,
61
+ "nauc_ndcg_at_1000_max": 0.765444,
62
+ "nauc_ndcg_at_1000_std": -0.140652,
63
+ "nauc_ndcg_at_1000_diff1": 0.89351,
64
+ "nauc_map_at_1_max": 0.715384,
65
+ "nauc_map_at_1_std": -0.153748,
66
+ "nauc_map_at_1_diff1": 0.899327,
67
+ "nauc_map_at_3_max": 0.759775,
68
+ "nauc_map_at_3_std": -0.145076,
69
+ "nauc_map_at_3_diff1": 0.891059,
70
+ "nauc_map_at_5_max": 0.763663,
71
+ "nauc_map_at_5_std": -0.142461,
72
+ "nauc_map_at_5_diff1": 0.893078,
73
+ "nauc_map_at_10_max": 0.760277,
74
+ "nauc_map_at_10_std": -0.14281,
75
+ "nauc_map_at_10_diff1": 0.893869,
76
+ "nauc_map_at_20_max": 0.759133,
77
+ "nauc_map_at_20_std": -0.143398,
78
+ "nauc_map_at_20_diff1": 0.893755,
79
+ "nauc_map_at_100_max": 0.758219,
80
+ "nauc_map_at_100_std": -0.144197,
81
+ "nauc_map_at_100_diff1": 0.893814,
82
+ "nauc_map_at_1000_max": 0.758178,
83
+ "nauc_map_at_1000_std": -0.144159,
84
+ "nauc_map_at_1000_diff1": 0.893822,
85
+ "nauc_recall_at_1_max": 0.715384,
86
+ "nauc_recall_at_1_std": -0.153748,
87
+ "nauc_recall_at_1_diff1": 0.899327,
88
+ "nauc_recall_at_3_max": 0.886372,
89
+ "nauc_recall_at_3_std": -0.085411,
90
+ "nauc_recall_at_3_diff1": 0.868253,
91
+ "nauc_recall_at_5_max": 0.962786,
92
+ "nauc_recall_at_5_std": -0.040735,
93
+ "nauc_recall_at_5_diff1": 0.881065,
94
+ "nauc_recall_at_10_max": 0.978942,
95
+ "nauc_recall_at_10_std": 0.003005,
96
+ "nauc_recall_at_10_diff1": 0.90083,
97
+ "nauc_recall_at_20_max": 0.991287,
98
+ "nauc_recall_at_20_std": 0.015147,
99
+ "nauc_recall_at_20_diff1": 0.892244,
100
+ "nauc_recall_at_100_max": 0.973861,
101
+ "nauc_recall_at_100_std": -0.408745,
102
+ "nauc_recall_at_100_diff1": 0.884699,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.715384,
107
+ "nauc_precision_at_1_std": -0.153748,
108
+ "nauc_precision_at_1_diff1": 0.899327,
109
+ "nauc_precision_at_3_max": 0.886372,
110
+ "nauc_precision_at_3_std": -0.085411,
111
+ "nauc_precision_at_3_diff1": 0.868253,
112
+ "nauc_precision_at_5_max": 0.962786,
113
+ "nauc_precision_at_5_std": -0.040735,
114
+ "nauc_precision_at_5_diff1": 0.881065,
115
+ "nauc_precision_at_10_max": 0.978942,
116
+ "nauc_precision_at_10_std": 0.003005,
117
+ "nauc_precision_at_10_diff1": 0.90083,
118
+ "nauc_precision_at_20_max": 0.991287,
119
+ "nauc_precision_at_20_std": 0.015147,
120
+ "nauc_precision_at_20_diff1": 0.892244,
121
+ "nauc_precision_at_100_max": 0.973861,
122
+ "nauc_precision_at_100_std": -0.408745,
123
+ "nauc_precision_at_100_diff1": 0.884699,
124
+ "nauc_precision_at_1000_max": 1.0,
125
+ "nauc_precision_at_1000_std": 1.0,
126
+ "nauc_precision_at_1000_diff1": 1.0,
127
+ "nauc_mrr_at_1_max": 0.715384,
128
+ "nauc_mrr_at_1_std": -0.153748,
129
+ "nauc_mrr_at_1_diff1": 0.899327,
130
+ "nauc_mrr_at_3_max": 0.759775,
131
+ "nauc_mrr_at_3_std": -0.145076,
132
+ "nauc_mrr_at_3_diff1": 0.891059,
133
+ "nauc_mrr_at_5_max": 0.763663,
134
+ "nauc_mrr_at_5_std": -0.142461,
135
+ "nauc_mrr_at_5_diff1": 0.893078,
136
+ "nauc_mrr_at_10_max": 0.760277,
137
+ "nauc_mrr_at_10_std": -0.14281,
138
+ "nauc_mrr_at_10_diff1": 0.893869,
139
+ "nauc_mrr_at_20_max": 0.759133,
140
+ "nauc_mrr_at_20_std": -0.143398,
141
+ "nauc_mrr_at_20_diff1": 0.893755,
142
+ "nauc_mrr_at_100_max": 0.758219,
143
+ "nauc_mrr_at_100_std": -0.144197,
144
+ "nauc_mrr_at_100_diff1": 0.893814,
145
+ "nauc_mrr_at_1000_max": 0.758178,
146
+ "nauc_mrr_at_1000_std": -0.144159,
147
+ "nauc_mrr_at_1000_diff1": 0.893822,
148
+ "main_score": 0.913733,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 5.1020729541778564,
157
+ "kg_co2_emissions": null
158
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/MTOPDomainClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d80d48c1eb48d3562165c59d59d0034df9fff0bf",
3
+ "task_name": "MTOPDomainClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.860661,
9
+ "f1": 0.84838,
10
+ "f1_weighted": 0.860317,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.873829,
14
+ "f1": 0.862083,
15
+ "f1_weighted": 0.873636
16
+ },
17
+ {
18
+ "accuracy": 0.839669,
19
+ "f1": 0.829691,
20
+ "f1_weighted": 0.839844
21
+ },
22
+ {
23
+ "accuracy": 0.854545,
24
+ "f1": 0.837422,
25
+ "f1_weighted": 0.854302
26
+ },
27
+ {
28
+ "accuracy": 0.869972,
29
+ "f1": 0.857268,
30
+ "f1_weighted": 0.868732
31
+ },
32
+ {
33
+ "accuracy": 0.872176,
34
+ "f1": 0.85866,
35
+ "f1_weighted": 0.871064
36
+ },
37
+ {
38
+ "accuracy": 0.857851,
39
+ "f1": 0.838354,
40
+ "f1_weighted": 0.855244
41
+ },
42
+ {
43
+ "accuracy": 0.863912,
44
+ "f1": 0.854162,
45
+ "f1_weighted": 0.863789
46
+ },
47
+ {
48
+ "accuracy": 0.842975,
49
+ "f1": 0.835944,
50
+ "f1_weighted": 0.844849
51
+ },
52
+ {
53
+ "accuracy": 0.860055,
54
+ "f1": 0.851354,
55
+ "f1_weighted": 0.860699
56
+ },
57
+ {
58
+ "accuracy": 0.871625,
59
+ "f1": 0.858866,
60
+ "f1_weighted": 0.871011
61
+ }
62
+ ],
63
+ "main_score": 0.860661,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.862468,
73
+ "f1": 0.847967,
74
+ "f1_weighted": 0.861675,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.873204,
78
+ "f1": 0.8603,
79
+ "f1_weighted": 0.872062
80
+ },
81
+ {
82
+ "accuracy": 0.840518,
83
+ "f1": 0.826207,
84
+ "f1_weighted": 0.840088
85
+ },
86
+ {
87
+ "accuracy": 0.861651,
88
+ "f1": 0.839463,
89
+ "f1_weighted": 0.860128
90
+ },
91
+ {
92
+ "accuracy": 0.872077,
93
+ "f1": 0.859522,
94
+ "f1_weighted": 0.871429
95
+ },
96
+ {
97
+ "accuracy": 0.871231,
98
+ "f1": 0.856149,
99
+ "f1_weighted": 0.869944
100
+ },
101
+ {
102
+ "accuracy": 0.857706,
103
+ "f1": 0.835122,
104
+ "f1_weighted": 0.854029
105
+ },
106
+ {
107
+ "accuracy": 0.86785,
108
+ "f1": 0.8532,
109
+ "f1_weighted": 0.867412
110
+ },
111
+ {
112
+ "accuracy": 0.84869,
113
+ "f1": 0.840454,
114
+ "f1_weighted": 0.850568
115
+ },
116
+ {
117
+ "accuracy": 0.857988,
118
+ "f1": 0.848332,
119
+ "f1_weighted": 0.85862
120
+ },
121
+ {
122
+ "accuracy": 0.873767,
123
+ "f1": 0.860924,
124
+ "f1_weighted": 0.872476
125
+ }
126
+ ],
127
+ "main_score": 0.862468,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 21.148167371749878,
136
+ "kg_co2_emissions": null
137
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/MTOPIntentClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba",
3
+ "task_name": "MTOPIntentClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.557906,
9
+ "f1": 0.350758,
10
+ "f1_weighted": 0.601537,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.548209,
14
+ "f1": 0.348813,
15
+ "f1_weighted": 0.596274
16
+ },
17
+ {
18
+ "accuracy": 0.560882,
19
+ "f1": 0.369328,
20
+ "f1_weighted": 0.608902
21
+ },
22
+ {
23
+ "accuracy": 0.560331,
24
+ "f1": 0.361627,
25
+ "f1_weighted": 0.610043
26
+ },
27
+ {
28
+ "accuracy": 0.550964,
29
+ "f1": 0.334527,
30
+ "f1_weighted": 0.585601
31
+ },
32
+ {
33
+ "accuracy": 0.577961,
34
+ "f1": 0.36162,
35
+ "f1_weighted": 0.620009
36
+ },
37
+ {
38
+ "accuracy": 0.53719,
39
+ "f1": 0.348759,
40
+ "f1_weighted": 0.567439
41
+ },
42
+ {
43
+ "accuracy": 0.577961,
44
+ "f1": 0.334588,
45
+ "f1_weighted": 0.618905
46
+ },
47
+ {
48
+ "accuracy": 0.572452,
49
+ "f1": 0.351694,
50
+ "f1_weighted": 0.619202
51
+ },
52
+ {
53
+ "accuracy": 0.521212,
54
+ "f1": 0.343009,
55
+ "f1_weighted": 0.569724
56
+ },
57
+ {
58
+ "accuracy": 0.571901,
59
+ "f1": 0.35361,
60
+ "f1_weighted": 0.619271
61
+ }
62
+ ],
63
+ "main_score": 0.557906,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.570668,
73
+ "f1": 0.358673,
74
+ "f1_weighted": 0.61582,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.572837,
78
+ "f1": 0.371875,
79
+ "f1_weighted": 0.619742
80
+ },
81
+ {
82
+ "accuracy": 0.559594,
83
+ "f1": 0.363088,
84
+ "f1_weighted": 0.611641
85
+ },
86
+ {
87
+ "accuracy": 0.57481,
88
+ "f1": 0.352673,
89
+ "f1_weighted": 0.624732
90
+ },
91
+ {
92
+ "accuracy": 0.551423,
93
+ "f1": 0.345272,
94
+ "f1_weighted": 0.591735
95
+ },
96
+ {
97
+ "accuracy": 0.582981,
98
+ "f1": 0.365224,
99
+ "f1_weighted": 0.628691
100
+ },
101
+ {
102
+ "accuracy": 0.538743,
103
+ "f1": 0.350986,
104
+ "f1_weighted": 0.570269
105
+ },
106
+ {
107
+ "accuracy": 0.595097,
108
+ "f1": 0.35422,
109
+ "f1_weighted": 0.638484
110
+ },
111
+ {
112
+ "accuracy": 0.600451,
113
+ "f1": 0.365976,
114
+ "f1_weighted": 0.648315
115
+ },
116
+ {
117
+ "accuracy": 0.53339,
118
+ "f1": 0.349534,
119
+ "f1_weighted": 0.58278
120
+ },
121
+ {
122
+ "accuracy": 0.597351,
123
+ "f1": 0.367881,
124
+ "f1_weighted": 0.641814
125
+ }
126
+ ],
127
+ "main_score": 0.570668,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 28.479371786117554,
136
+ "kg_co2_emissions": null
137
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/PawsXPairClassification.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "8a04d940a42cd40658986fdd8e3da561533a3646",
3
+ "task_name": "PawsXPairClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "similarity_accuracy": 0.5845,
9
+ "similarity_accuracy_threshold": 0.992572,
10
+ "similarity_f1": 0.624245,
11
+ "similarity_f1_threshold": 0.929772,
12
+ "similarity_precision": 0.457769,
13
+ "similarity_recall": 0.981006,
14
+ "similarity_ap": 0.541329,
15
+ "cosine_accuracy": 0.5845,
16
+ "cosine_accuracy_threshold": 0.992572,
17
+ "cosine_f1": 0.624245,
18
+ "cosine_f1_threshold": 0.929772,
19
+ "cosine_precision": 0.457769,
20
+ "cosine_recall": 0.981006,
21
+ "cosine_ap": 0.538849,
22
+ "manhattan_accuracy": 0.5855,
23
+ "manhattan_accuracy_threshold": 1.898788,
24
+ "manhattan_f1": 0.624108,
25
+ "manhattan_f1_threshold": 5.749293,
26
+ "manhattan_precision": 0.458355,
27
+ "manhattan_recall": 0.977654,
28
+ "manhattan_ap": 0.539746,
29
+ "euclidean_accuracy": 0.5845,
30
+ "euclidean_accuracy_threshold": 0.121885,
31
+ "euclidean_f1": 0.624245,
32
+ "euclidean_f1_threshold": 0.374775,
33
+ "euclidean_precision": 0.457769,
34
+ "euclidean_recall": 0.981006,
35
+ "euclidean_ap": 0.538848,
36
+ "dot_accuracy": 0.5845,
37
+ "dot_accuracy_threshold": 0.992572,
38
+ "dot_f1": 0.624245,
39
+ "dot_f1_threshold": 0.929772,
40
+ "dot_precision": 0.457769,
41
+ "dot_recall": 0.981006,
42
+ "dot_ap": 0.53894,
43
+ "max_accuracy": 0.5855,
44
+ "max_f1": 0.624245,
45
+ "max_precision": 0.458355,
46
+ "max_recall": 0.981006,
47
+ "max_ap": 0.541329,
48
+ "main_score": 0.541329,
49
+ "hf_subset": "de",
50
+ "languages": [
51
+ "deu-Latn"
52
+ ]
53
+ }
54
+ ],
55
+ "validation": [
56
+ {
57
+ "similarity_accuracy": 0.5965,
58
+ "similarity_accuracy_threshold": 0.999701,
59
+ "similarity_f1": 0.599567,
60
+ "similarity_f1_threshold": 0.875372,
61
+ "similarity_precision": 0.42813,
62
+ "similarity_recall": 1.0,
63
+ "similarity_ap": 0.487436,
64
+ "cosine_accuracy": 0.5965,
65
+ "cosine_accuracy_threshold": 0.999701,
66
+ "cosine_f1": 0.599567,
67
+ "cosine_f1_threshold": 0.875372,
68
+ "cosine_precision": 0.42813,
69
+ "cosine_recall": 1.0,
70
+ "cosine_ap": 0.482428,
71
+ "manhattan_accuracy": 0.5965,
72
+ "manhattan_accuracy_threshold": 0.386895,
73
+ "manhattan_f1": 0.599783,
74
+ "manhattan_f1_threshold": 7.41747,
75
+ "manhattan_precision": 0.428351,
76
+ "manhattan_recall": 1.0,
77
+ "manhattan_ap": 0.483387,
78
+ "euclidean_accuracy": 0.5965,
79
+ "euclidean_accuracy_threshold": 0.024388,
80
+ "euclidean_f1": 0.599567,
81
+ "euclidean_f1_threshold": 0.498898,
82
+ "euclidean_precision": 0.42813,
83
+ "euclidean_recall": 1.0,
84
+ "euclidean_ap": 0.482428,
85
+ "dot_accuracy": 0.5985,
86
+ "dot_accuracy_threshold": 1.0,
87
+ "dot_f1": 0.599567,
88
+ "dot_f1_threshold": 0.875372,
89
+ "dot_precision": 0.42813,
90
+ "dot_recall": 1.0,
91
+ "dot_ap": 0.491956,
92
+ "max_accuracy": 0.5985,
93
+ "max_f1": 0.599783,
94
+ "max_precision": 0.428351,
95
+ "max_recall": 1.0,
96
+ "max_ap": 0.491956,
97
+ "main_score": 0.491956,
98
+ "hf_subset": "de",
99
+ "languages": [
100
+ "deu-Latn"
101
+ ]
102
+ }
103
+ ]
104
+ },
105
+ "evaluation_time": 5.304517507553101,
106
+ "kg_co2_emissions": null
107
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/XMarket.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "dfe57acff5b62c23732a7b7d3e3fb84ff501708b",
3
+ "task_name": "XMarket",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.1516,
9
+ "ndcg_at_3": 0.14916,
10
+ "ndcg_at_5": 0.15081,
11
+ "ndcg_at_10": 0.15546,
12
+ "ndcg_at_20": 0.16259,
13
+ "ndcg_at_100": 0.18567,
14
+ "ndcg_at_1000": 0.22237,
15
+ "map_at_1": 0.04396,
16
+ "map_at_3": 0.06778,
17
+ "map_at_5": 0.07767,
18
+ "map_at_10": 0.08817,
19
+ "map_at_20": 0.09588,
20
+ "map_at_100": 0.10507,
21
+ "map_at_1000": 0.10942,
22
+ "recall_at_1": 0.04396,
23
+ "recall_at_3": 0.08557,
24
+ "recall_at_5": 0.11032,
25
+ "recall_at_10": 0.14697,
26
+ "recall_at_20": 0.18567,
27
+ "recall_at_100": 0.28615,
28
+ "recall_at_1000": 0.458,
29
+ "precision_at_1": 0.1516,
30
+ "precision_at_3": 0.12336,
31
+ "precision_at_5": 0.10711,
32
+ "precision_at_10": 0.08449,
33
+ "precision_at_20": 0.06329,
34
+ "precision_at_100": 0.02857,
35
+ "precision_at_1000": 0.00812,
36
+ "mrr_at_1": 0.151598,
37
+ "mrr_at_3": 0.196598,
38
+ "mrr_at_5": 0.208104,
39
+ "mrr_at_10": 0.216678,
40
+ "mrr_at_20": 0.220609,
41
+ "mrr_at_100": 0.223962,
42
+ "mrr_at_1000": 0.224626,
43
+ "nauc_ndcg_at_1_max": 0.203044,
44
+ "nauc_ndcg_at_1_std": 0.240929,
45
+ "nauc_ndcg_at_1_diff1": 0.167017,
46
+ "nauc_ndcg_at_3_max": 0.192645,
47
+ "nauc_ndcg_at_3_std": 0.256287,
48
+ "nauc_ndcg_at_3_diff1": 0.13872,
49
+ "nauc_ndcg_at_5_max": 0.187711,
50
+ "nauc_ndcg_at_5_std": 0.260568,
51
+ "nauc_ndcg_at_5_diff1": 0.136172,
52
+ "nauc_ndcg_at_10_max": 0.177508,
53
+ "nauc_ndcg_at_10_std": 0.259459,
54
+ "nauc_ndcg_at_10_diff1": 0.133588,
55
+ "nauc_ndcg_at_20_max": 0.169441,
56
+ "nauc_ndcg_at_20_std": 0.257989,
57
+ "nauc_ndcg_at_20_diff1": 0.13327,
58
+ "nauc_ndcg_at_100_max": 0.160949,
59
+ "nauc_ndcg_at_100_std": 0.266419,
60
+ "nauc_ndcg_at_100_diff1": 0.128889,
61
+ "nauc_ndcg_at_1000_max": 0.166864,
62
+ "nauc_ndcg_at_1000_std": 0.296552,
63
+ "nauc_ndcg_at_1000_diff1": 0.121614,
64
+ "nauc_map_at_1_max": 0.089683,
65
+ "nauc_map_at_1_std": 0.041819,
66
+ "nauc_map_at_1_diff1": 0.302405,
67
+ "nauc_map_at_3_max": 0.09996,
68
+ "nauc_map_at_3_std": 0.07114,
69
+ "nauc_map_at_3_diff1": 0.231627,
70
+ "nauc_map_at_5_max": 0.110681,
71
+ "nauc_map_at_5_std": 0.09655,
72
+ "nauc_map_at_5_diff1": 0.208783,
73
+ "nauc_map_at_10_max": 0.125103,
74
+ "nauc_map_at_10_std": 0.132127,
75
+ "nauc_map_at_10_diff1": 0.189581,
76
+ "nauc_map_at_20_max": 0.139756,
77
+ "nauc_map_at_20_std": 0.163407,
78
+ "nauc_map_at_20_diff1": 0.176591,
79
+ "nauc_map_at_100_max": 0.156901,
80
+ "nauc_map_at_100_std": 0.201285,
81
+ "nauc_map_at_100_diff1": 0.163176,
82
+ "nauc_map_at_1000_max": 0.165915,
83
+ "nauc_map_at_1000_std": 0.220585,
84
+ "nauc_map_at_1000_diff1": 0.158211,
85
+ "nauc_recall_at_1_max": 0.089683,
86
+ "nauc_recall_at_1_std": 0.041819,
87
+ "nauc_recall_at_1_diff1": 0.302405,
88
+ "nauc_recall_at_3_max": 0.074723,
89
+ "nauc_recall_at_3_std": 0.064755,
90
+ "nauc_recall_at_3_diff1": 0.19245,
91
+ "nauc_recall_at_5_max": 0.076601,
92
+ "nauc_recall_at_5_std": 0.091138,
93
+ "nauc_recall_at_5_diff1": 0.160936,
94
+ "nauc_recall_at_10_max": 0.076469,
95
+ "nauc_recall_at_10_std": 0.126776,
96
+ "nauc_recall_at_10_diff1": 0.126922,
97
+ "nauc_recall_at_20_max": 0.082321,
98
+ "nauc_recall_at_20_std": 0.161643,
99
+ "nauc_recall_at_20_diff1": 0.113748,
100
+ "nauc_recall_at_100_max": 0.079585,
101
+ "nauc_recall_at_100_std": 0.207257,
102
+ "nauc_recall_at_100_diff1": 0.095673,
103
+ "nauc_recall_at_1000_max": 0.066772,
104
+ "nauc_recall_at_1000_std": 0.249238,
105
+ "nauc_recall_at_1000_diff1": 0.084073,
106
+ "nauc_precision_at_1_max": 0.203044,
107
+ "nauc_precision_at_1_std": 0.240929,
108
+ "nauc_precision_at_1_diff1": 0.167017,
109
+ "nauc_precision_at_3_max": 0.222879,
110
+ "nauc_precision_at_3_std": 0.317938,
111
+ "nauc_precision_at_3_diff1": 0.085462,
112
+ "nauc_precision_at_5_max": 0.236988,
113
+ "nauc_precision_at_5_std": 0.366147,
114
+ "nauc_precision_at_5_diff1": 0.054348,
115
+ "nauc_precision_at_10_max": 0.253212,
116
+ "nauc_precision_at_10_std": 0.418504,
117
+ "nauc_precision_at_10_diff1": 0.023799,
118
+ "nauc_precision_at_20_max": 0.266308,
119
+ "nauc_precision_at_20_std": 0.456984,
120
+ "nauc_precision_at_20_diff1": 0.000257,
121
+ "nauc_precision_at_100_max": 0.264595,
122
+ "nauc_precision_at_100_std": 0.489513,
123
+ "nauc_precision_at_100_diff1": -0.022013,
124
+ "nauc_precision_at_1000_max": 0.219897,
125
+ "nauc_precision_at_1000_std": 0.412952,
126
+ "nauc_precision_at_1000_diff1": -0.056858,
127
+ "nauc_mrr_at_1_max": 0.203044,
128
+ "nauc_mrr_at_1_std": 0.240929,
129
+ "nauc_mrr_at_1_diff1": 0.167017,
130
+ "nauc_mrr_at_3_max": 0.190864,
131
+ "nauc_mrr_at_3_std": 0.258584,
132
+ "nauc_mrr_at_3_diff1": 0.145435,
133
+ "nauc_mrr_at_5_max": 0.184383,
134
+ "nauc_mrr_at_5_std": 0.260719,
135
+ "nauc_mrr_at_5_diff1": 0.139316,
136
+ "nauc_mrr_at_10_max": 0.182958,
137
+ "nauc_mrr_at_10_std": 0.264431,
138
+ "nauc_mrr_at_10_diff1": 0.137192,
139
+ "nauc_mrr_at_20_max": 0.182317,
140
+ "nauc_mrr_at_20_std": 0.264007,
141
+ "nauc_mrr_at_20_diff1": 0.13717,
142
+ "nauc_mrr_at_100_max": 0.181592,
143
+ "nauc_mrr_at_100_std": 0.264616,
144
+ "nauc_mrr_at_100_diff1": 0.13694,
145
+ "nauc_mrr_at_1000_max": 0.181664,
146
+ "nauc_mrr_at_1000_std": 0.264636,
147
+ "nauc_mrr_at_1000_diff1": 0.136895,
148
+ "main_score": 0.15546,
149
+ "hf_subset": "de",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 48.49527955055237,
157
+ "kg_co2_emissions": null
158
+ }
results/intfloat__multilingual-e5-small/fd1525a9fd15316a2d503bf26ab031a61d056e98/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "intfloat/multilingual-e5-small", "revision": "fd1525a9fd15316a2d503bf26ab031a61d056e98", "release_date": "2024-02-08", "languages": ["afr_Latn", "amh_Latn", "ara_Latn", "asm_Latn", "aze_Latn", "bel_Latn", "bul_Latn", "ben_Latn", "ben_Beng", "bre_Latn", "bos_Latn", "cat_Latn", "ces_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "ell_Latn", "eng_Latn", "epo_Latn", "spa_Latn", "est_Latn", "eus_Latn", "fas_Latn", "fin_Latn", "fra_Latn", "fry_Latn", "gle_Latn", "gla_Latn", "glg_Latn", "guj_Latn", "hau_Latn", "heb_Latn", "hin_Latn", "hin_Deva", "hrv_Latn", "hun_Latn", "hye_Latn", "ind_Latn", "isl_Latn", "ita_Latn", "jpn_Latn", "jav_Latn", "kat_Latn", "kaz_Latn", "khm_Latn", "kan_Latn", "kor_Latn", "kur_Latn", "kir_Latn", "lat_Latn", "lao_Latn", "lit_Latn", "lav_Latn", "mlg_Latn", "mkd_Latn", "mal_Latn", "mon_Latn", "mar_Latn", "msa_Latn", "mya_Latn", "nep_Latn", "nld_Latn", "nob_Latn", "orm_Latn", "ori_Latn", "pan_Latn", "pol_Latn", "pus_Latn", "por_Latn", "ron_Latn", "rus_Latn", "san_Latn", "snd_Latn", "sin_Latn", "slk_Latn", "slv_Latn", "som_Latn", "sqi_Latn", "srp_Latn", "sun_Latn", "swe_Latn", "swa_Latn", "tam_Latn", "tam_Taml", "tel_Latn", "tel_Telu", "tha_Latn", "tgl_Latn", "tur_Latn", "uig_Latn", "ukr_Latn", "urd_Latn", "urd_Arab", "uzb_Latn", "vie_Latn", "xho_Latn", "yid_Latn", "zho_Hant", "zho_Hans"], "n_parameters": 118000000, "memory_usage": null, "max_tokens": 512.0, "embed_dim": 384, "license": "mit", "open_weights": true, "public_training_data": null, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/intfloat/multilingual-e5-small", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": null, "adapted_from": null, "superseded_by": null, "loader": "sentence_transformers_loader"}
 
1
+ {"name": "intfloat/multilingual-e5-small", "revision": "fd1525a9fd15316a2d503bf26ab031a61d056e98", "release_date": "2024-02-08", "languages": ["afr_Latn", "amh_Latn", "ara_Latn", "asm_Latn", "aze_Latn", "bel_Latn", "bul_Latn", "ben_Latn", "ben_Beng", "bre_Latn", "bos_Latn", "cat_Latn", "ces_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "ell_Latn", "eng_Latn", "epo_Latn", "spa_Latn", "est_Latn", "eus_Latn", "fas_Latn", "fin_Latn", "fra_Latn", "fry_Latn", "gle_Latn", "gla_Latn", "glg_Latn", "guj_Latn", "hau_Latn", "heb_Latn", "hin_Latn", "hin_Deva", "hrv_Latn", "hun_Latn", "hye_Latn", "ind_Latn", "isl_Latn", "ita_Latn", "jpn_Latn", "jav_Latn", "kat_Latn", "kaz_Latn", "khm_Latn", "kan_Latn", "kor_Latn", "kur_Latn", "kir_Latn", "lat_Latn", "lao_Latn", "lit_Latn", "lav_Latn", "mlg_Latn", "mkd_Latn", "mal_Latn", "mon_Latn", "mar_Latn", "msa_Latn", "mya_Latn", "nep_Latn", "nld_Latn", "nob_Latn", "orm_Latn", "ori_Latn", "pan_Latn", "pol_Latn", "pus_Latn", "por_Latn", "ron_Latn", "rus_Latn", "san_Latn", "snd_Latn", "sin_Latn", "slk_Latn", "slv_Latn", "som_Latn", "sqi_Latn", "srp_Latn", "sun_Latn", "swe_Latn", "swa_Latn", "tam_Latn", "tam_Taml", "tel_Latn", "tel_Telu", "tha_Latn", "tgl_Latn", "tur_Latn", "uig_Latn", "ukr_Latn", "urd_Latn", "urd_Arab", "uzb_Latn", "vie_Latn", "xho_Latn", "yid_Latn", "zho_Hant", "zho_Hans"], "n_parameters": 118000000, "max_tokens": 512.0, "embed_dim": 384, "license": "mit", "open_weights": true, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/intfloat/multilingual-e5-small", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": {"MSMARCO": ["train"], "MSMARCOHardNegatives": ["train"], "NanoMSMARCORetrieval": ["train"], "MSMARCO-PL": ["train"], "NQ": ["train"], "NQHardNegatives": ["train"], "NanoNQRetrieval": ["train"], "NQ-PL": ["train"]}, "adapted_from": null, "superseded_by": null, "loader": "sentence_transformers_loader"}
results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/AmazonCounterfactualClassification.json ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
3
+ "task_name": "AmazonCounterfactualClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.695279,
9
+ "f1": 0.673793,
10
+ "f1_weighted": 0.706517,
11
+ "ap": 0.803739,
12
+ "ap_weighted": 0.803739,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.699571,
16
+ "f1": 0.681843,
17
+ "f1_weighted": 0.711497,
18
+ "ap": 0.812194,
19
+ "ap_weighted": 0.812194
20
+ },
21
+ {
22
+ "accuracy": 0.714592,
23
+ "f1": 0.678048,
24
+ "f1_weighted": 0.720877,
25
+ "ap": 0.791447,
26
+ "ap_weighted": 0.791447
27
+ },
28
+ {
29
+ "accuracy": 0.714592,
30
+ "f1": 0.690102,
31
+ "f1_weighted": 0.7245,
32
+ "ap": 0.80844,
33
+ "ap_weighted": 0.80844
34
+ },
35
+ {
36
+ "accuracy": 0.693133,
37
+ "f1": 0.680404,
38
+ "f1_weighted": 0.705588,
39
+ "ap": 0.819962,
40
+ "ap_weighted": 0.819962
41
+ },
42
+ {
43
+ "accuracy": 0.706009,
44
+ "f1": 0.683993,
45
+ "f1_weighted": 0.716927,
46
+ "ap": 0.807738,
47
+ "ap_weighted": 0.807738
48
+ },
49
+ {
50
+ "accuracy": 0.716738,
51
+ "f1": 0.70189,
52
+ "f1_weighted": 0.72816,
53
+ "ap": 0.829804,
54
+ "ap_weighted": 0.829804
55
+ },
56
+ {
57
+ "accuracy": 0.695279,
58
+ "f1": 0.673637,
59
+ "f1_weighted": 0.706821,
60
+ "ap": 0.802038,
61
+ "ap_weighted": 0.802038
62
+ },
63
+ {
64
+ "accuracy": 0.60515,
65
+ "f1": 0.581851,
66
+ "f1_weighted": 0.620824,
67
+ "ap": 0.747753,
68
+ "ap_weighted": 0.747753
69
+ },
70
+ {
71
+ "accuracy": 0.712446,
72
+ "f1": 0.688179,
73
+ "f1_weighted": 0.722527,
74
+ "ap": 0.807575,
75
+ "ap_weighted": 0.807575
76
+ },
77
+ {
78
+ "accuracy": 0.695279,
79
+ "f1": 0.677983,
80
+ "f1_weighted": 0.70745,
81
+ "ap": 0.810439,
82
+ "ap_weighted": 0.810439
83
+ }
84
+ ],
85
+ "main_score": 0.695279,
86
+ "hf_subset": "de",
87
+ "languages": [
88
+ "deu-Latn"
89
+ ]
90
+ }
91
+ ],
92
+ "test": [
93
+ {
94
+ "accuracy": 0.687045,
95
+ "f1": 0.665641,
96
+ "f1_weighted": 0.698448,
97
+ "ap": 0.796931,
98
+ "ap_weighted": 0.796931,
99
+ "scores_per_experiment": [
100
+ {
101
+ "accuracy": 0.684154,
102
+ "f1": 0.664326,
103
+ "f1_weighted": 0.696295,
104
+ "ap": 0.796805,
105
+ "ap_weighted": 0.796805
106
+ },
107
+ {
108
+ "accuracy": 0.702355,
109
+ "f1": 0.668622,
110
+ "f1_weighted": 0.710053,
111
+ "ap": 0.786676,
112
+ "ap_weighted": 0.786676
113
+ },
114
+ {
115
+ "accuracy": 0.708779,
116
+ "f1": 0.682651,
117
+ "f1_weighted": 0.718334,
118
+ "ap": 0.800707,
119
+ "ap_weighted": 0.800707
120
+ },
121
+ {
122
+ "accuracy": 0.667024,
123
+ "f1": 0.654031,
124
+ "f1_weighted": 0.680303,
125
+ "ap": 0.800195,
126
+ "ap_weighted": 0.800195
127
+ },
128
+ {
129
+ "accuracy": 0.701285,
130
+ "f1": 0.681461,
131
+ "f1_weighted": 0.7126,
132
+ "ap": 0.807184,
133
+ "ap_weighted": 0.807184
134
+ },
135
+ {
136
+ "accuracy": 0.721627,
137
+ "f1": 0.702642,
138
+ "f1_weighted": 0.732085,
139
+ "ap": 0.821279,
140
+ "ap_weighted": 0.821279
141
+ },
142
+ {
143
+ "accuracy": 0.638116,
144
+ "f1": 0.612073,
145
+ "f1_weighted": 0.65146,
146
+ "ap": 0.76081,
147
+ "ap_weighted": 0.76081
148
+ },
149
+ {
150
+ "accuracy": 0.639186,
151
+ "f1": 0.618195,
152
+ "f1_weighted": 0.653276,
153
+ "ap": 0.768243,
154
+ "ap_weighted": 0.768243
155
+ },
156
+ {
157
+ "accuracy": 0.721627,
158
+ "f1": 0.698632,
159
+ "f1_weighted": 0.731253,
160
+ "ap": 0.813397,
161
+ "ap_weighted": 0.813397
162
+ },
163
+ {
164
+ "accuracy": 0.686296,
165
+ "f1": 0.673772,
166
+ "f1_weighted": 0.698819,
167
+ "ap": 0.814014,
168
+ "ap_weighted": 0.814014
169
+ }
170
+ ],
171
+ "main_score": 0.687045,
172
+ "hf_subset": "de",
173
+ "languages": [
174
+ "deu-Latn"
175
+ ]
176
+ }
177
+ ]
178
+ },
179
+ "evaluation_time": 24.15538454055786,
180
+ "kg_co2_emissions": null
181
+ }
results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/AmazonReviewsClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1399c76144fd37290681b995c656ef9b2e06e26d",
3
+ "task_name": "AmazonReviewsClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.33596,
9
+ "f1": 0.33313,
10
+ "f1_weighted": 0.33313,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.3858,
14
+ "f1": 0.384151,
15
+ "f1_weighted": 0.384151
16
+ },
17
+ {
18
+ "accuracy": 0.32,
19
+ "f1": 0.316271,
20
+ "f1_weighted": 0.316271
21
+ },
22
+ {
23
+ "accuracy": 0.3356,
24
+ "f1": 0.328853,
25
+ "f1_weighted": 0.328853
26
+ },
27
+ {
28
+ "accuracy": 0.3396,
29
+ "f1": 0.334589,
30
+ "f1_weighted": 0.334589
31
+ },
32
+ {
33
+ "accuracy": 0.3334,
34
+ "f1": 0.331795,
35
+ "f1_weighted": 0.331795
36
+ },
37
+ {
38
+ "accuracy": 0.3446,
39
+ "f1": 0.343729,
40
+ "f1_weighted": 0.343729
41
+ },
42
+ {
43
+ "accuracy": 0.2794,
44
+ "f1": 0.281941,
45
+ "f1_weighted": 0.281941
46
+ },
47
+ {
48
+ "accuracy": 0.3552,
49
+ "f1": 0.354249,
50
+ "f1_weighted": 0.354249
51
+ },
52
+ {
53
+ "accuracy": 0.328,
54
+ "f1": 0.325996,
55
+ "f1_weighted": 0.325996
56
+ },
57
+ {
58
+ "accuracy": 0.338,
59
+ "f1": 0.329723,
60
+ "f1_weighted": 0.329723
61
+ }
62
+ ],
63
+ "main_score": 0.33596,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.33394,
73
+ "f1": 0.331978,
74
+ "f1_weighted": 0.331978,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.3748,
78
+ "f1": 0.374319,
79
+ "f1_weighted": 0.374319
80
+ },
81
+ {
82
+ "accuracy": 0.325,
83
+ "f1": 0.323455,
84
+ "f1_weighted": 0.323455
85
+ },
86
+ {
87
+ "accuracy": 0.3342,
88
+ "f1": 0.329901,
89
+ "f1_weighted": 0.329901
90
+ },
91
+ {
92
+ "accuracy": 0.3386,
93
+ "f1": 0.335603,
94
+ "f1_weighted": 0.335603
95
+ },
96
+ {
97
+ "accuracy": 0.3326,
98
+ "f1": 0.332394,
99
+ "f1_weighted": 0.332394
100
+ },
101
+ {
102
+ "accuracy": 0.3364,
103
+ "f1": 0.337795,
104
+ "f1_weighted": 0.337795
105
+ },
106
+ {
107
+ "accuracy": 0.2918,
108
+ "f1": 0.294337,
109
+ "f1_weighted": 0.294337
110
+ },
111
+ {
112
+ "accuracy": 0.3466,
113
+ "f1": 0.34317,
114
+ "f1_weighted": 0.34317
115
+ },
116
+ {
117
+ "accuracy": 0.3226,
118
+ "f1": 0.319688,
119
+ "f1_weighted": 0.319688
120
+ },
121
+ {
122
+ "accuracy": 0.3368,
123
+ "f1": 0.329121,
124
+ "f1_weighted": 0.329121
125
+ }
126
+ ],
127
+ "main_score": 0.33394,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 37.28035569190979,
136
+ "kg_co2_emissions": null
137
+ }
results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/MTOPDomainClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d80d48c1eb48d3562165c59d59d0034df9fff0bf",
3
+ "task_name": "MTOPDomainClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.778292,
9
+ "f1": 0.765488,
10
+ "f1_weighted": 0.779069,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.772452,
14
+ "f1": 0.760727,
15
+ "f1_weighted": 0.770152
16
+ },
17
+ {
18
+ "accuracy": 0.780165,
19
+ "f1": 0.764418,
20
+ "f1_weighted": 0.77965
21
+ },
22
+ {
23
+ "accuracy": 0.77741,
24
+ "f1": 0.755123,
25
+ "f1_weighted": 0.777198
26
+ },
27
+ {
28
+ "accuracy": 0.783471,
29
+ "f1": 0.772344,
30
+ "f1_weighted": 0.783796
31
+ },
32
+ {
33
+ "accuracy": 0.76584,
34
+ "f1": 0.753206,
35
+ "f1_weighted": 0.767035
36
+ },
37
+ {
38
+ "accuracy": 0.792287,
39
+ "f1": 0.77434,
40
+ "f1_weighted": 0.791956
41
+ },
42
+ {
43
+ "accuracy": 0.800551,
44
+ "f1": 0.786026,
45
+ "f1_weighted": 0.800623
46
+ },
47
+ {
48
+ "accuracy": 0.768044,
49
+ "f1": 0.761453,
50
+ "f1_weighted": 0.772465
51
+ },
52
+ {
53
+ "accuracy": 0.774105,
54
+ "f1": 0.770192,
55
+ "f1_weighted": 0.777779
56
+ },
57
+ {
58
+ "accuracy": 0.768595,
59
+ "f1": 0.757055,
60
+ "f1_weighted": 0.770035
61
+ }
62
+ ],
63
+ "main_score": 0.778292,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.786081,
73
+ "f1": 0.771171,
74
+ "f1_weighted": 0.786473,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.786137,
78
+ "f1": 0.77482,
79
+ "f1_weighted": 0.784724
80
+ },
81
+ {
82
+ "accuracy": 0.766413,
83
+ "f1": 0.747166,
84
+ "f1_weighted": 0.76688
85
+ },
86
+ {
87
+ "accuracy": 0.782192,
88
+ "f1": 0.757721,
89
+ "f1_weighted": 0.781977
90
+ },
91
+ {
92
+ "accuracy": 0.797971,
93
+ "f1": 0.782559,
94
+ "f1_weighted": 0.798152
95
+ },
96
+ {
97
+ "accuracy": 0.778247,
98
+ "f1": 0.764502,
99
+ "f1_weighted": 0.778991
100
+ },
101
+ {
102
+ "accuracy": 0.801071,
103
+ "f1": 0.784254,
104
+ "f1_weighted": 0.802288
105
+ },
106
+ {
107
+ "accuracy": 0.802761,
108
+ "f1": 0.78556,
109
+ "f1_weighted": 0.803478
110
+ },
111
+ {
112
+ "accuracy": 0.783601,
113
+ "f1": 0.77684,
114
+ "f1_weighted": 0.786652
115
+ },
116
+ {
117
+ "accuracy": 0.786419,
118
+ "f1": 0.776707,
119
+ "f1_weighted": 0.786679
120
+ },
121
+ {
122
+ "accuracy": 0.775993,
123
+ "f1": 0.761579,
124
+ "f1_weighted": 0.77491
125
+ }
126
+ ],
127
+ "main_score": 0.786081,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 22.63331937789917,
136
+ "kg_co2_emissions": null
137
+ }
results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/MTOPIntentClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba",
3
+ "task_name": "MTOPIntentClassification",
4
+ "mteb_version": "1.29.10",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.546832,
9
+ "f1": 0.356526,
10
+ "f1_weighted": 0.593777,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.555923,
14
+ "f1": 0.369012,
15
+ "f1_weighted": 0.599311
16
+ },
17
+ {
18
+ "accuracy": 0.551515,
19
+ "f1": 0.365264,
20
+ "f1_weighted": 0.601706
21
+ },
22
+ {
23
+ "accuracy": 0.567493,
24
+ "f1": 0.376636,
25
+ "f1_weighted": 0.610419
26
+ },
27
+ {
28
+ "accuracy": 0.557576,
29
+ "f1": 0.352619,
30
+ "f1_weighted": 0.607518
31
+ },
32
+ {
33
+ "accuracy": 0.536088,
34
+ "f1": 0.35328,
35
+ "f1_weighted": 0.58046
36
+ },
37
+ {
38
+ "accuracy": 0.524518,
39
+ "f1": 0.35448,
40
+ "f1_weighted": 0.574988
41
+ },
42
+ {
43
+ "accuracy": 0.543251,
44
+ "f1": 0.34267,
45
+ "f1_weighted": 0.593026
46
+ },
47
+ {
48
+ "accuracy": 0.552066,
49
+ "f1": 0.367481,
50
+ "f1_weighted": 0.596148
51
+ },
52
+ {
53
+ "accuracy": 0.537741,
54
+ "f1": 0.350639,
55
+ "f1_weighted": 0.587173
56
+ },
57
+ {
58
+ "accuracy": 0.542149,
59
+ "f1": 0.333179,
60
+ "f1_weighted": 0.587019
61
+ }
62
+ ],
63
+ "main_score": 0.546832,
64
+ "hf_subset": "de",
65
+ "languages": [
66
+ "deu-Latn"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.552578,
73
+ "f1": 0.379394,
74
+ "f1_weighted": 0.598045,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.56523,
78
+ "f1": 0.38411,
79
+ "f1_weighted": 0.60955
80
+ },
81
+ {
82
+ "accuracy": 0.54466,
83
+ "f1": 0.379038,
84
+ "f1_weighted": 0.593165
85
+ },
86
+ {
87
+ "accuracy": 0.560158,
88
+ "f1": 0.393551,
89
+ "f1_weighted": 0.595091
90
+ },
91
+ {
92
+ "accuracy": 0.55255,
93
+ "f1": 0.36243,
94
+ "f1_weighted": 0.605802
95
+ },
96
+ {
97
+ "accuracy": 0.550859,
98
+ "f1": 0.383485,
99
+ "f1_weighted": 0.595259
100
+ },
101
+ {
102
+ "accuracy": 0.522119,
103
+ "f1": 0.360199,
104
+ "f1_weighted": 0.568394
105
+ },
106
+ {
107
+ "accuracy": 0.559031,
108
+ "f1": 0.393272,
109
+ "f1_weighted": 0.605839
110
+ },
111
+ {
112
+ "accuracy": 0.551423,
113
+ "f1": 0.381976,
114
+ "f1_weighted": 0.59388
115
+ },
116
+ {
117
+ "accuracy": 0.560158,
118
+ "f1": 0.378374,
119
+ "f1_weighted": 0.607985
120
+ },
121
+ {
122
+ "accuracy": 0.559594,
123
+ "f1": 0.377507,
124
+ "f1_weighted": 0.605487
125
+ }
126
+ ],
127
+ "main_score": 0.552578,
128
+ "hf_subset": "de",
129
+ "languages": [
130
+ "deu-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 26.755773067474365,
136
+ "kg_co2_emissions": null
137
+ }
results/shibing624__text2vec-base-multilingual/6633dc49e554de7105458f8f2e96445c6598e9d1/model_meta.json CHANGED
@@ -1 +1 @@
1
- {"name": "shibing624/text2vec-base-multilingual", "revision": "6633dc49e554de7105458f8f2e96445c6598e9d1", "release_date": "2023-06-22", "languages": ["zho_Hans", "eng_Latn", "deu_Latn", "fra_Latn", "ita_Latn", "nld_Latn", "por_Latn", "pol_Latn", "rus_Cyrl"], "n_parameters": 117654272, "memory_usage": null, "max_tokens": 512.0, "embed_dim": 384, "license": "apache-2.0", "open_weights": true, "public_training_data": true, "public_training_code": null, "framework": ["PyTorch", "Sentence Transformers"], "reference": "https://huggingface.co/shibing624/text2vec-base-multilingual", "similarity_fn_name": "cosine", "use_instructions": null, "training_datasets": {"shibing624/nli-zh-all": ["train"]}, "adapted_from": "sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2", "superseded_by": null, "loader": null}
 
1
+ {"name": "shibing624/text2vec-base-multilingual", "revision": "6633dc49e554de7105458f8f2e96445c6598e9d1", "release_date": "2023-06-22", "languages": ["deu-Latn", "eng-Latn", "spa-Latn", "fra-Latn", "ita-Latn", "nld-Latn", "pol-Latn", "por-Latn", "rus-Cyrl", "zho-Hans"], "n_parameters": 118000000, "max_tokens": 256.0, "embed_dim": 384, "license": "apache-2.0", "open_weights": true, "public_training_code": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/shibing624/text2vec-base-chinese-paraphrase", "similarity_fn_name": "cosine", "use_instructions": false, "training_datasets": {}, "adapted_from": "sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2", "superseded_by": null, "loader": null}