IlyasMoutawwakil HF staff commited on
Commit
e02a1e3
·
verified ·
1 Parent(s): eaf773d

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -6,17 +6,19 @@
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
- "library": "transformers",
10
  "model": "google-bert/bert-base-uncased",
11
- "processor": "google-bert/bert-base-uncased",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
- "model_kwargs": {},
18
- "processor_kwargs": {},
19
- "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
@@ -86,7 +88,7 @@
86
  "optimum_benchmark_commit": null,
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
- "accelerate_version": "0.30.0",
90
  "accelerate_commit": null,
91
  "diffusers_version": "0.27.2",
92
  "diffusers_commit": null,
@@ -102,7 +104,7 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 866.246656,
106
  "max_global_vram": 1229.45536,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 591.396864,
@@ -110,180 +112,187 @@
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 149,
114
- "total": 1.0044498558044437,
115
- "mean": 0.00674127420002982,
116
- "stdev": 0.00024499234659723505,
117
- "p50": 0.006795263767242431,
118
- "p90": 0.0069777409553527835,
119
- "p95": 0.00704778242111206,
120
- "p99": 0.007468851203918457,
121
  "values": [
122
- 0.007434239864349365,
123
- 0.006978559970855713,
124
- 0.007173120021820068,
125
- 0.007058432102203369,
126
- 0.007072768211364746,
127
- 0.007023615837097168,
128
- 0.007021567821502686,
129
- 0.006977536201477051,
130
- 0.007021567821502686,
131
- 0.006884352207183838,
132
- 0.006872064113616944,
133
- 0.006855679988861084,
134
- 0.006873087882995605,
135
- 0.006795263767242431,
136
- 0.006754303932189941,
137
- 0.006833151817321777,
138
- 0.006866943836212158,
139
- 0.0068853759765625,
140
- 0.006851583957672119,
141
- 0.0069027838706970214,
142
- 0.006892543792724609,
143
- 0.006841343879699707,
144
- 0.0068577280044555666,
145
- 0.006822912216186523,
146
- 0.0069253120422363285,
147
- 0.007029759883880615,
148
- 0.007021567821502686,
149
- 0.006944767951965332,
150
- 0.006875135898590088,
151
- 0.006823935985565186,
152
- 0.006845439910888672,
153
- 0.006864895820617676,
154
- 0.006854656219482422,
155
- 0.0067983360290527345,
156
- 0.006751232147216797,
157
- 0.0067717118263244626,
158
- 0.006812672138214112,
159
- 0.006847487926483154,
160
- 0.006819839954376221,
161
- 0.006812672138214112,
162
- 0.0068618240356445315,
163
- 0.006876160144805908,
164
- 0.006904831886291504,
165
- 0.006948863983154297,
166
- 0.006918144226074219,
167
- 0.006884352207183838,
168
- 0.006821887969970703,
169
- 0.006847487926483154,
170
- 0.006871039867401123,
171
- 0.006880256175994873,
172
- 0.0068280320167541505,
173
- 0.0068577280044555666,
174
- 0.006882304191589355,
175
- 0.006729728221893311,
176
- 0.006830080032348633,
177
- 0.006808576107025147,
178
- 0.006767615795135498,
179
- 0.006959104061126709,
180
- 0.0068884482383728025,
181
- 0.006758399963378906,
182
- 0.006800384044647217,
183
- 0.006809599876403808,
184
- 0.006815743923187256,
185
- 0.006767615795135498,
186
- 0.0075008001327514645,
187
- 0.00692633581161499,
188
- 0.006948863983154297,
189
- 0.0068986878395080565,
190
- 0.006977536201477051,
191
- 0.006905856132507324,
192
- 0.007031807899475098,
193
- 0.006810624122619629,
194
- 0.006892543792724609,
195
- 0.006879231929779053,
196
- 0.006841343879699707,
197
- 0.006848512172698974,
198
- 0.006846464157104492,
199
- 0.006792191982269287,
200
- 0.006677567958831787,
201
- 0.006620160102844238,
202
- 0.00679423999786377,
203
- 0.006659071922302246,
204
- 0.006495232105255127,
205
- 0.006543360233306885,
206
  0.0065781760215759275,
207
- 0.0066447358131408694,
208
- 0.006759424209594727,
209
- 0.006744063854217529,
210
- 0.006841343879699707,
211
- 0.007374847888946533,
212
- 0.007998464107513427,
213
- 0.0073062400817871095,
214
- 0.0064849920272827145,
215
- 0.00652288007736206,
216
- 0.006500351905822754,
217
- 0.006461440086364746,
218
- 0.006479872226715088,
219
- 0.006461440086364746,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
220
  0.006511616230010986,
221
- 0.006523903846740723,
222
- 0.006486015796661377,
223
- 0.00653004789352417,
224
- 0.006551551818847656,
225
- 0.0064880638122558594,
226
- 0.006557695865631104,
227
- 0.0065136637687683106,
228
- 0.006561791896820069,
229
- 0.006482944011688232,
230
- 0.006503424167633057,
231
- 0.006470655918121338,
232
- 0.006479872226715088,
233
- 0.006490111827850342,
234
- 0.006498303890228272,
235
- 0.006471680164337158,
236
- 0.0065136637687683106,
237
- 0.006494207859039307,
238
- 0.006489088058471679,
239
- 0.006491136074066162,
240
- 0.006505472183227539,
241
- 0.006470655918121338,
242
- 0.006471680164337158,
243
- 0.006453248023986816,
244
- 0.006497280120849609,
245
- 0.006558720111846924,
246
- 0.006470655918121338,
247
- 0.0064767999649047855,
248
- 0.006532095909118653,
249
- 0.006464511871337891,
250
- 0.006471680164337158,
251
- 0.006466559886932373,
252
- 0.006494207859039307,
253
- 0.006487040042877197,
254
- 0.006471680164337158,
255
- 0.006479872226715088,
256
- 0.006486015796661377,
257
  0.006435840129852295,
258
- 0.0064767999649047855,
259
- 0.006460415840148926,
260
- 0.006496255874633789,
261
- 0.006501376152038574,
262
- 0.006496255874633789,
263
- 0.006489088058471679,
264
- 0.006499328136444092,
265
- 0.006491136074066162,
266
- 0.006854656219482422,
267
- 0.00659660816192627,
268
- 0.006492159843444824,
269
- 0.0065484800338745115,
270
- 0.006526976108551025
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
271
  ]
272
  },
273
  "throughput": {
274
  "unit": "samples/s",
275
- "value": 148.33990879581438
276
  },
277
  "energy": {
278
  "unit": "kWh",
279
- "cpu": 7.197752043052957e-08,
280
- "ram": 3.930865874281153e-08,
281
- "gpu": 1.3703714666667818e-07,
282
- "total": 2.483233258400193e-07
283
  },
284
  "efficiency": {
285
  "unit": "samples/kWh",
286
- "value": 4027007.9204892884
287
  }
288
  }
289
  }
 
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
 
9
  "model": "google-bert/bert-base-uncased",
10
+ "library": "transformers",
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": false
21
+ },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
 
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.40.2",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.30.1",
92
  "accelerate_commit": null,
93
  "diffusers_version": "0.27.2",
94
  "diffusers_commit": null,
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 866.28352,
108
  "max_global_vram": 1229.45536,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 591.396864,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 156,
116
+ "total": 1.0038590397834781,
117
+ "mean": 0.006434993844765882,
118
+ "stdev": 0.00017473140964596726,
119
+ "p50": 0.006380032062530518,
120
+ "p90": 0.0067164161205291745,
121
+ "p95": 0.006760704159736634,
122
+ "p99": 0.006924646592140197,
123
  "values": [
124
+ 0.007416831970214844,
125
+ 0.006780928134918213,
126
+ 0.006971392154693603,
127
+ 0.0066375679969787596,
128
+ 0.00657919979095459,
129
+ 0.0067348480224609375,
130
+ 0.006628352165222168,
131
+ 0.006549503803253174,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
132
  0.0065781760215759275,
133
+ 0.006560768127441406,
134
+ 0.006495232105255127,
135
+ 0.006545407772064209,
136
+ 0.006719488143920899,
137
+ 0.006708223819732666,
138
+ 0.006649856090545654,
139
+ 0.0066007041931152345,
140
+ 0.006643712043762207,
141
+ 0.006421504020690918,
142
+ 0.006421504020690918,
143
+ 0.006467584133148193,
144
+ 0.006493184089660644,
145
+ 0.006456319808959961,
146
+ 0.00677785587310791,
147
+ 0.006670335769653321,
148
+ 0.006649856090545654,
149
+ 0.006730751991271973,
150
+ 0.00679423999786377,
151
+ 0.00669593620300293,
152
+ 0.006713344097137451,
153
+ 0.006743040084838867,
154
+ 0.006677504062652588,
155
+ 0.006719488143920899,
156
+ 0.006478847980499268,
157
+ 0.006419456005096436,
158
+ 0.006450175762176514,
159
+ 0.0063569917678833006,
160
+ 0.006319104194641113,
161
+ 0.006358016014099121,
162
+ 0.006750207901000976,
163
+ 0.006436863899230957,
164
+ 0.006418432235717773,
165
+ 0.006725632190704346,
166
+ 0.00688640022277832,
167
+ 0.00677785587310791,
168
+ 0.006755328178405762,
169
+ 0.006776832103729248,
170
  0.006511616230010986,
171
+ 0.006432767868041992,
172
+ 0.0064440321922302245,
173
+ 0.00638156795501709,
174
+ 0.006404096126556396,
175
+ 0.006465536117553711,
176
+ 0.006421504020690918,
177
+ 0.006402048110961914,
178
+ 0.0064245758056640625,
179
+ 0.00643174409866333,
180
+ 0.006368256092071533,
181
+ 0.0064204797744750975,
182
+ 0.006408192157745361,
183
+ 0.00637337589263916,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
184
  0.006435840129852295,
185
+ 0.0064542717933654785,
186
+ 0.006341631889343262,
187
+ 0.006354944229125976,
188
+ 0.00638976001739502,
189
+ 0.006455296039581298,
190
+ 0.006402080059051514,
191
+ 0.006367231845855713,
192
+ 0.006440959930419922,
193
+ 0.006397952079772949,
194
+ 0.006500351905822754,
195
+ 0.006481919765472412,
196
+ 0.006459392070770263,
197
+ 0.006468607902526856,
198
+ 0.006502399921417237,
199
+ 0.006400000095367431,
200
+ 0.00642252779006958,
201
+ 0.0064287037849426265,
202
+ 0.0063498239517211915,
203
+ 0.006388735771179199,
204
+ 0.00637337589263916,
205
+ 0.006352896213531494,
206
+ 0.006313983917236328,
207
+ 0.00623308801651001,
208
+ 0.006271999835968017,
209
+ 0.006371327877044678,
210
+ 0.006232063770294189,
211
+ 0.006072319984436035,
212
+ 0.00608460807800293,
213
+ 0.0062873601913452145,
214
+ 0.006355967998504639,
215
+ 0.006392831802368164,
216
+ 0.006302720069885254,
217
+ 0.006384640216827392,
218
+ 0.006378496170043945,
219
+ 0.006352896213531494,
220
+ 0.006281216144561768,
221
+ 0.006335487842559814,
222
+ 0.006333439826965332,
223
+ 0.006310912132263183,
224
+ 0.006308864116668702,
225
+ 0.006299647808074951,
226
+ 0.006326272010803223,
227
+ 0.006326272010803223,
228
+ 0.006372352123260498,
229
+ 0.006338560104370118,
230
+ 0.006322175979614258,
231
+ 0.006313983917236328,
232
+ 0.006323200225830078,
233
+ 0.0063272957801818845,
234
+ 0.006299647808074951,
235
+ 0.006322175979614258,
236
+ 0.00643891191482544,
237
+ 0.006352896213531494,
238
+ 0.006302720069885254,
239
+ 0.0062863359451293946,
240
+ 0.006347775936126709,
241
+ 0.006311935901641846,
242
+ 0.006295551776885986,
243
+ 0.00632422399520874,
244
+ 0.006289408206939697,
245
+ 0.006311935901641846,
246
+ 0.006299647808074951,
247
+ 0.006303743839263916,
248
+ 0.0064225602149963375,
249
+ 0.00628223991394043,
250
+ 0.006322175979614258,
251
+ 0.006285312175750732,
252
+ 0.0063907837867736815,
253
+ 0.006309887886047363,
254
+ 0.006316031932830811,
255
+ 0.006319104194641113,
256
+ 0.0063201279640197755,
257
+ 0.0063201279640197755,
258
+ 0.006302720069885254,
259
+ 0.00638976001739502,
260
+ 0.0063211522102355954,
261
+ 0.006277120113372803,
262
+ 0.0063272957801818845,
263
+ 0.006307839870452881,
264
+ 0.006319104194641113,
265
+ 0.0063498239517211915,
266
+ 0.0063539199829101565,
267
+ 0.006301695823669433,
268
+ 0.0063272957801818845,
269
+ 0.006330368041992188,
270
+ 0.006343679904937744,
271
+ 0.006369279861450195,
272
+ 0.006376448154449463,
273
+ 0.006333439826965332,
274
+ 0.006333439826965332,
275
+ 0.006427648067474365,
276
+ 0.006310912132263183,
277
+ 0.006310912132263183,
278
+ 0.006301695823669433,
279
+ 0.006312960147857666
280
  ]
281
  },
282
  "throughput": {
283
  "unit": "samples/s",
284
+ "value": 155.4003040443284
285
  },
286
  "energy": {
287
  "unit": "kWh",
288
+ "cpu": 7.476132912372559e-08,
289
+ "ram": 4.0869070684459316e-08,
290
+ "gpu": 1.4018180327043992e-07,
291
+ "total": 2.5581220307862484e-07
292
  },
293
  "efficiency": {
294
  "unit": "samples/kWh",
295
+ "value": 3909117.657270815
296
  }
297
  }
298
  }