IlyasMoutawwakil HF staff commited on
Commit
42808e9
·
verified ·
1 Parent(s): e03eee2

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "forward": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 866.041856,
6
+ "max_global_vram": 1229.45536,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 591.396864,
9
+ "max_allocated": 449.269248
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 155,
14
+ "total": 1.002450977802276,
15
+ "mean": 0.0064674256632404935,
16
+ "stdev": 0.00019750743205881847,
17
+ "p50": 0.006493184089660644,
18
+ "p90": 0.006580838203430176,
19
+ "p95": 0.006742732906341552,
20
+ "p99": 0.007255674724578858,
21
+ "values": [
22
+ 0.007309311866760254,
23
+ 0.006625279903411865,
24
+ 0.006788095951080322,
25
+ 0.0067276802062988285,
26
+ 0.00677785587310791,
27
+ 0.006856704235076904,
28
+ 0.006940671920776367,
29
+ 0.007209983825683594,
30
+ 0.006408192157745361,
31
+ 0.006195199966430664,
32
+ 0.006196224212646485,
33
+ 0.006168575763702393,
34
+ 0.006195199966430664,
35
+ 0.006160384178161621,
36
+ 0.006171648025512695,
37
+ 0.006162432193756104,
38
+ 0.006174719810485839,
39
+ 0.006160384178161621,
40
+ 0.006202367782592774,
41
+ 0.006140927791595459,
42
+ 0.00616755199432373,
43
+ 0.00628326416015625,
44
+ 0.006170623779296875,
45
+ 0.006187007904052734,
46
+ 0.00617574405670166,
47
+ 0.006168575763702393,
48
+ 0.006144000053405762,
49
+ 0.006236159801483154,
50
+ 0.006176767826080322,
51
+ 0.006243328094482422,
52
+ 0.006169600009918213,
53
+ 0.0062269439697265625,
54
+ 0.006196224212646485,
55
+ 0.006207520008087158,
56
+ 0.006182911872863769,
57
+ 0.006154240131378174,
58
+ 0.006174719810485839,
59
+ 0.006189055919647217,
60
+ 0.006191103935241699,
61
+ 0.006149119853973388,
62
+ 0.006277120113372803,
63
+ 0.006719488143920899,
64
+ 0.006461440086364746,
65
+ 0.006491136074066162,
66
+ 0.006457344055175781,
67
+ 0.006525951862335205,
68
+ 0.006495232105255127,
69
+ 0.006493184089660644,
70
+ 0.006493184089660644,
71
+ 0.006469632148742676,
72
+ 0.006491136074066162,
73
+ 0.006497280120849609,
74
+ 0.006467584133148193,
75
+ 0.006497280120849609,
76
+ 0.006504447937011719,
77
+ 0.006491136074066162,
78
+ 0.006500351905822754,
79
+ 0.006498303890228272,
80
+ 0.006474751949310303,
81
+ 0.006489088058471679,
82
+ 0.006460415840148926,
83
+ 0.0064839677810668945,
84
+ 0.006501376152038574,
85
+ 0.006474751949310303,
86
+ 0.0065064959526062015,
87
+ 0.006545407772064209,
88
+ 0.006502399921417237,
89
+ 0.006491136074066162,
90
+ 0.006464511871337891,
91
+ 0.006503424167633057,
92
+ 0.0065064959526062015,
93
+ 0.006495232105255127,
94
+ 0.006605823993682861,
95
+ 0.00653004789352417,
96
+ 0.006479872226715088,
97
+ 0.006504447937011719,
98
+ 0.006493184089660644,
99
+ 0.006510591983795166,
100
+ 0.006501376152038574,
101
+ 0.006449151992797852,
102
+ 0.006516736030578613,
103
+ 0.006502399921417237,
104
+ 0.006490111827850342,
105
+ 0.006560768127441406,
106
+ 0.006478847980499268,
107
+ 0.006493184089660644,
108
+ 0.006502399921417237,
109
+ 0.006455296039581298,
110
+ 0.006487040042877197,
111
+ 0.0064767999649047855,
112
+ 0.006482944011688232,
113
+ 0.0065146880149841305,
114
+ 0.006464511871337891,
115
+ 0.006516736030578613,
116
+ 0.0064778242111206055,
117
+ 0.006469632148742676,
118
+ 0.0064737281799316405,
119
+ 0.006469632148742676,
120
+ 0.006479872226715088,
121
+ 0.006512639999389648,
122
+ 0.006583295822143555,
123
+ 0.0065064959526062015,
124
+ 0.006512639999389648,
125
+ 0.00653926420211792,
126
+ 0.00653107213973999,
127
+ 0.0065382399559021,
128
+ 0.006501376152038574,
129
+ 0.006574079990386963,
130
+ 0.006510591983795166,
131
+ 0.006525951862335205,
132
+ 0.006500351905822754,
133
+ 0.006501376152038574,
134
+ 0.00653004789352417,
135
+ 0.006499328136444092,
136
+ 0.006529024124145508,
137
+ 0.0065771517753601075,
138
+ 0.006499328136444092,
139
+ 0.006472703933715821,
140
+ 0.006491136074066162,
141
+ 0.006448128223419189,
142
+ 0.006478847980499268,
143
+ 0.006474751949310303,
144
+ 0.006497280120849609,
145
+ 0.006499328136444092,
146
+ 0.006472703933715821,
147
+ 0.006470655918121338,
148
+ 0.0065177597999572755,
149
+ 0.0064737281799316405,
150
+ 0.006494207859039307,
151
+ 0.006499328136444092,
152
+ 0.006496255874633789,
153
+ 0.006642687797546387,
154
+ 0.006604800224304199,
155
+ 0.006497280120849609,
156
+ 0.006510591983795166,
157
+ 0.006435840129852295,
158
+ 0.00643993616104126,
159
+ 0.006425600051879882,
160
+ 0.006412288188934326,
161
+ 0.006460415840148926,
162
+ 0.006385663986206055,
163
+ 0.006462463855743408,
164
+ 0.0065064959526062015,
165
+ 0.006667263984680176,
166
+ 0.007445504188537597,
167
+ 0.00694374418258667,
168
+ 0.006558720111846924,
169
+ 0.006471680164337158,
170
+ 0.006562816143035889,
171
+ 0.006504447937011719,
172
+ 0.006575104236602783,
173
+ 0.006555647850036621,
174
+ 0.0065484800338745115,
175
+ 0.006540287971496582,
176
+ 0.006549503803253174
177
+ ]
178
+ },
179
+ "throughput": {
180
+ "unit": "samples/s",
181
+ "value": 154.6210272943364
182
+ },
183
+ "energy": {
184
+ "unit": "kWh",
185
+ "cpu": 7.693178871626496e-08,
186
+ "ram": 4.193464474918719e-08,
187
+ "gpu": 1.4653595952257972e-07,
188
+ "total": 2.654023929880319e-07
189
+ },
190
+ "efficiency": {
191
+ "unit": "samples/kWh",
192
+ "value": 3767863.5401192266
193
+ }
194
+ }
195
+ }