IlyasMoutawwakil HF staff commited on
Commit
9a442b0
·
verified ·
1 Parent(s): 27b6794

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 787.755008,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.4037890625,
117
- "mean": 7.4037890625,
118
  "stdev": 0.0,
119
- "p50": 7.4037890625,
120
- "p90": 7.4037890625,
121
- "p95": 7.4037890625,
122
- "p99": 7.4037890625,
123
  "values": [
124
- 7.4037890625
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 1.0830540388884202e-06,
131
- "ram": 5.604532750178463e-07,
132
- "gpu": 1.722779156000144e-06,
133
- "total": 3.36628646990641e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1077.69856,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
@@ -145,182 +145,192 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 151,
149
- "total": 0.9990293745994568,
150
- "mean": 0.006616088573506336,
151
- "stdev": 0.00017019617599665937,
152
- "p50": 0.006652927875518798,
153
- "p90": 0.006758399963378906,
154
- "p95": 0.006814719915390014,
155
- "p99": 0.007146512031555176,
156
  "values": [
157
- 0.007186431884765625,
158
- 0.007106592178344726,
159
- 0.007102464199066162,
160
- 0.006967296123504638,
161
- 0.007012383937835693,
162
- 0.0069253120422363285,
163
- 0.007488512039184571,
164
- 0.006628352165222168,
165
- 0.006489088058471679,
166
- 0.006503424167633057,
167
- 0.006446080207824707,
168
- 0.006779903888702392,
169
- 0.006821887969970703,
170
- 0.006688767910003662,
171
- 0.006729728221893311,
172
- 0.006727583885192871,
173
- 0.006491136074066162,
174
- 0.0064778242111206055,
175
- 0.006479872226715088,
176
- 0.006441984176635742,
177
- 0.0064880638122558594,
178
- 0.0067573761940002445,
179
- 0.006457344055175781,
180
- 0.00648089599609375,
181
- 0.0064327998161315915,
182
- 0.006452223777770996,
183
- 0.006460415840148926,
184
- 0.006527999877929688,
185
- 0.006457344055175781,
186
- 0.0064102401733398436,
187
- 0.006448128223419189,
188
- 0.00657919979095459,
189
- 0.006651904106140137,
190
- 0.006716383934020996,
191
- 0.00667852783203125,
192
- 0.006677504062652588,
193
- 0.006749184131622315,
194
- 0.006699007987976074,
195
- 0.006694911956787109,
196
  0.006692863941192627,
197
- 0.006718463897705078,
198
- 0.0067010560035705566,
199
- 0.006684671878814697,
200
- 0.0066979842185974124,
201
- 0.006683648109436035,
202
- 0.0064358081817626955,
203
- 0.006451200008392334,
204
- 0.006419456005096436,
205
- 0.006464511871337891,
206
- 0.0064542717933654785,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
207
  0.006412288188934326,
208
- 0.0066826238632202144,
209
- 0.0066938881874084475,
210
- 0.00667952013015747,
211
- 0.006700992107391357,
212
- 0.006710271835327148,
213
- 0.006687744140625,
214
- 0.006703104019165039,
215
- 0.006725632190704346,
216
- 0.0067051520347595215,
217
- 0.0067051520347595215,
218
- 0.006706175804138184,
219
- 0.006676479816436768,
220
- 0.006702079772949219,
221
- 0.006706175804138184,
222
- 0.006685696125030518,
223
- 0.006796288013458252,
224
- 0.006710207939147949,
225
- 0.0067051520347595215,
226
- 0.006649856090545654,
227
- 0.006716415882110595,
228
- 0.006719488143920899,
229
- 0.006659071922302246,
230
- 0.006732800006866455,
231
- 0.0066979842185974124,
232
- 0.006746047973632812,
233
- 0.006758399963378906,
234
- 0.006683648109436035,
235
- 0.006644703865051269,
236
- 0.00648089599609375,
237
- 0.006642687797546387,
238
- 0.006715392112731934,
239
- 0.0066826558113098145,
240
- 0.006699007987976074,
241
- 0.006657023906707763,
242
- 0.006677504062652588,
243
- 0.0066109437942504885,
244
- 0.006565887928009034,
245
- 0.0064245758056640625,
246
- 0.006481887817382813,
247
- 0.006527999877929688,
248
- 0.006502399921417237,
249
- 0.006490111827850342,
250
- 0.006441984176635742,
251
- 0.006456319808959961,
252
- 0.006408192157745361,
253
- 0.006652927875518798,
254
- 0.006670335769653321,
255
- 0.006464511871337891,
256
- 0.006500351905822754,
257
- 0.006441984176635742,
258
- 0.006427648067474365,
259
- 0.006511616230010986,
260
- 0.006636544227600098,
261
- 0.0065484800338745115,
262
- 0.0067358717918396,
263
- 0.0068055038452148435,
264
- 0.006752319812774658,
265
- 0.006807551860809326,
266
- 0.0067870721817016606,
267
- 0.006753215789794922,
268
- 0.0067870721817016606,
269
- 0.006761472225189209,
270
- 0.006669312000274658,
271
- 0.006731776237487793,
272
- 0.006744063854217529,
273
- 0.006680575847625733,
274
- 0.006708223819732666,
275
- 0.0067358717918396,
276
- 0.006706175804138184,
277
- 0.006656000137329102,
278
- 0.006707200050354004,
279
- 0.006665215969085693,
280
- 0.0064471039772033695,
281
- 0.006455232143402099,
282
- 0.006426623821258545,
283
- 0.006470655918121338,
284
- 0.0064839677810668945,
285
- 0.006472703933715821,
286
- 0.006453248023986816,
287
- 0.006429696083068847,
288
- 0.0064245758056640625,
289
- 0.006491136074066162,
290
- 0.006405119895935059,
291
- 0.006461440086364746,
292
- 0.0064440321922302245,
293
- 0.006428671836853027,
294
- 0.006499328136444092,
295
- 0.006499328136444092,
296
- 0.006474751949310303,
297
- 0.006456319808959961,
298
- 0.006435840129852295,
299
- 0.006457344055175781,
300
- 0.006466559886932373,
301
- 0.0064143362045288085,
302
- 0.006452223777770996,
303
- 0.0064204797744750975,
304
- 0.006462463855743408,
305
- 0.006502399921417237,
306
- 0.006419456005096436,
307
- 0.00643891191482544
308
  ]
309
  },
310
  "throughput": {
311
  "unit": "samples/s",
312
- "value": 151.14670683286045
313
  },
314
  "energy": {
315
  "unit": "kWh",
316
- "cpu": 7.961012520041163e-08,
317
- "ram": 4.342889963197727e-08,
318
- "gpu": 1.4644120589262048e-07,
319
- "total": 2.6948023072500937e-07
320
  },
321
  "efficiency": {
322
  "unit": "samples/kWh",
323
- "value": 3710847.349765142
324
  }
325
  }
326
  }
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 763.838464,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.40290966796875,
117
+ "mean": 7.40290966796875,
118
  "stdev": 0.0,
119
+ "p50": 7.40290966796875,
120
+ "p90": 7.40290966796875,
121
+ "p95": 7.40290966796875,
122
+ "p99": 7.40290966796875,
123
  "values": [
124
+ 7.40290966796875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 1.115731651388864e-06,
131
+ "ram": 5.65891498107513e-07,
132
+ "gpu": 1.669168002000048e-06,
133
+ "total": 3.3507911514964248e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 957.083648,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 161,
149
+ "total": 0.999686176776886,
150
+ "mean": 0.006209230911657677,
151
+ "stdev": 0.00018290817969804406,
152
+ "p50": 0.006120448112487793,
153
+ "p90": 0.0064133119583129885,
154
+ "p95": 0.006498303890228272,
155
+ "p99": 0.006715596961975098,
156
  "values": [
157
+ 0.006667263984680176,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
158
  0.006692863941192627,
159
+ 0.006691904067993164,
160
+ 0.006700032234191895,
161
+ 0.006738944053649902,
162
+ 0.006633471965789795,
163
+ 0.007146495819091797,
164
+ 0.006336512088775635,
165
+ 0.006160384178161621,
166
+ 0.006181888103485108,
167
+ 0.00617574405670166,
168
+ 0.006146048069000244,
169
+ 0.0061634559631347655,
170
+ 0.00611740779876709,
171
+ 0.006139904022216797,
172
+ 0.006112256050109863,
173
+ 0.006157311916351318,
174
+ 0.006146048069000244,
175
+ 0.006110208034515381,
176
+ 0.006132736206054688,
177
+ 0.006082560062408447,
178
+ 0.006141952037811279,
179
+ 0.006082592010498047,
180
+ 0.006115327835083008,
181
+ 0.00608358383178711,
182
+ 0.006120448112487793,
183
+ 0.006089727878570556,
184
+ 0.006123519897460937,
185
+ 0.006085631847381592,
186
+ 0.006112256050109863,
187
+ 0.0060999679565429685,
188
+ 0.006083648204803466,
189
+ 0.006053887844085694,
190
+ 0.00601087999343872,
191
+ 0.006033408164978027,
192
+ 0.006001664161682129,
193
+ 0.006031360149383545,
194
+ 0.006012928009033203,
195
+ 0.006034431934356689,
196
+ 0.0059955201148986816,
197
+ 0.006055935859680176,
198
+ 0.005997568130493164,
199
+ 0.006038527965545654,
200
+ 0.006012928009033203,
201
+ 0.006047743797302246,
202
+ 0.006014944076538086,
203
+ 0.006049791812896729,
204
+ 0.006023168087005615,
205
+ 0.006041600227355957,
206
+ 0.006081535816192627,
207
+ 0.0061859841346740725,
208
+ 0.006114304065704346,
209
+ 0.00614188814163208,
210
+ 0.006104063987731933,
211
+ 0.006106112003326416,
212
+ 0.006105088233947754,
213
+ 0.006105088233947754,
214
+ 0.006139904022216797,
215
+ 0.006097951889038086,
216
+ 0.006124544143676758,
217
+ 0.006103040218353272,
218
+ 0.006119359970092773,
219
+ 0.006061056137084961,
220
+ 0.006121471881866455,
221
+ 0.006059999942779541,
222
+ 0.006102015972137451,
223
+ 0.006095871925354004,
224
+ 0.006118400096893311,
225
+ 0.006095871925354004,
226
+ 0.006124544143676758,
227
+ 0.006103040218353272,
228
+ 0.006091775894165039,
229
+ 0.006090752124786377,
230
+ 0.00606822395324707,
231
+ 0.006107135772705078,
232
+ 0.006072319984436035,
233
+ 0.006109183788299561,
234
+ 0.006072319984436035,
235
+ 0.006094848155975342,
236
+ 0.006082560062408447,
237
+ 0.006121471881866455,
238
+ 0.006056960105895996,
239
+ 0.0061296639442443845,
240
+ 0.006118400096893311,
241
+ 0.006138879776000977,
242
+ 0.006113279819488526,
243
+ 0.00608358383178711,
244
+ 0.0060928001403808595,
245
+ 0.006096896171569824,
246
+ 0.0061224961280822755,
247
+ 0.0060590081214904785,
248
+ 0.0060999679565429685,
249
+ 0.006160384178161621,
250
+ 0.006159359931945801,
251
+ 0.00608460807800293,
252
+ 0.006079487800598145,
253
+ 0.006080512046813965,
254
+ 0.006110208034515381,
255
+ 0.006081535816192627,
256
+ 0.006082560062408447,
257
+ 0.006079487800598145,
258
+ 0.006098944187164307,
259
+ 0.006065184116363525,
260
+ 0.006062079906463623,
261
+ 0.006061056137084961,
262
+ 0.006066239833831787,
263
+ 0.006085663795471191,
264
+ 0.006061056137084961,
265
+ 0.006096896171569824,
266
+ 0.006060031890869141,
267
+ 0.006097919940948486,
268
+ 0.006045695781707764,
269
+ 0.006552544116973877,
270
+ 0.006445055961608887,
271
+ 0.006455296039581298,
272
+ 0.006498303890228272,
273
+ 0.006377471923828125,
274
+ 0.006445055961608887,
275
+ 0.0064133119583129885,
276
+ 0.006404032230377197,
277
+ 0.006421504020690918,
278
+ 0.006341631889343262,
279
  0.006412288188934326,
280
+ 0.006398975849151611,
281
+ 0.00637440013885498,
282
+ 0.006481919765472412,
283
+ 0.00638156795501709,
284
+ 0.006421504020690918,
285
+ 0.006404096126556396,
286
+ 0.006351871967315674,
287
+ 0.006360064029693604,
288
+ 0.0063621439933776855,
289
+ 0.00638976001739502,
290
+ 0.006370304107666015,
291
+ 0.006318079948425293,
292
+ 0.00635590410232544,
293
+ 0.00632425594329834,
294
+ 0.006364160060882569,
295
+ 0.006369279861450195,
296
+ 0.006329343795776367,
297
+ 0.006355967998504639,
298
+ 0.006307839870452881,
299
+ 0.0063539199829101565,
300
+ 0.0063272957801818845,
301
+ 0.006331391811370849,
302
+ 0.00634771203994751,
303
+ 0.0063170561790466305,
304
+ 0.006359039783477783,
305
+ 0.006338560104370118,
306
+ 0.006347775936126709,
307
+ 0.00637440013885498,
308
+ 0.006334464073181153,
309
+ 0.00637440013885498,
310
+ 0.00636624002456665,
311
+ 0.006360064029693604,
312
+ 0.0063794879913330075,
313
+ 0.006310912132263183,
314
+ 0.006367263793945313,
315
+ 0.006326272010803223,
316
+ 0.0063836159706115725,
317
+ 0.0064174079895019534
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
318
  ]
319
  },
320
  "throughput": {
321
  "unit": "samples/s",
322
+ "value": 161.0505413999864
323
  },
324
  "energy": {
325
  "unit": "kWh",
326
+ "cpu": 7.579152883492722e-08,
327
+ "ram": 4.1428711470426856e-08,
328
+ "gpu": 1.362191464840761e-07,
329
+ "total": 2.5343938678943023e-07
330
  },
331
  "efficiency": {
332
  "unit": "samples/kWh",
333
+ "value": 3945716.6175628756
334
  }
335
  }
336
  }