IlyasMoutawwakil HF staff commited on
Commit
a1e3114
·
verified ·
1 Parent(s): 44700da

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -6,17 +6,19 @@
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
- "library": "transformers",
10
  "model": "FacebookAI/roberta-base",
11
- "processor": "FacebookAI/roberta-base",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
- "model_kwargs": {},
18
- "processor_kwargs": {},
19
- "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
@@ -71,7 +73,7 @@
71
  "environment": {
72
  "cpu": " AMD EPYC 7R32",
73
  "cpu_count": 16,
74
- "cpu_ram_mb": 66697.29792,
75
  "system": "Linux",
76
  "machine": "x86_64",
77
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
@@ -86,7 +88,7 @@
86
  "optimum_benchmark_commit": null,
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
- "accelerate_version": "0.30.0",
90
  "accelerate_commit": null,
91
  "diffusers_version": "0.27.2",
92
  "diffusers_commit": null,
@@ -102,7 +104,7 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 891.621376,
106
  "max_global_vram": 1195.900928,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 555.74528,
@@ -110,177 +112,167 @@
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 146,
114
- "total": 0.9972845778465275,
115
- "mean": 0.006830716286620049,
116
- "stdev": 0.0001990599915854344,
117
- "p50": 0.006783967971801758,
118
- "p90": 0.006865440130233765,
119
- "p95": 0.0073021438121795654,
120
- "p99": 0.00782781438827515,
121
  "values": [
122
- 0.008047616004943848,
123
- 0.0076267518997192385,
 
 
 
 
 
 
 
 
 
 
 
124
  0.0074291200637817386,
125
- 0.007386015892028809,
126
- 0.007299071788787842,
127
- 0.007346176147460938,
128
- 0.007336959838867187,
129
- 0.00799232006072998,
130
- 0.006978559970855713,
131
- 0.006864960193634033,
132
- 0.0067900800704956054,
133
- 0.0068269438743591305,
134
- 0.006820799827575684,
135
- 0.006787007808685303,
136
- 0.006843391895294189,
137
- 0.006874112129211426,
138
- 0.006866943836212158,
139
- 0.00678604793548584,
140
- 0.0068269438743591305,
141
- 0.006811647891998291,
142
- 0.006774784088134766,
143
- 0.006823935985565186,
144
- 0.006814720153808594,
145
- 0.006822912216186523,
146
- 0.006793216228485107,
147
- 0.006838272094726563,
148
- 0.006816768169403077,
149
- 0.0067870721817016606,
150
- 0.006806528091430664,
151
- 0.0068197760581970215,
152
- 0.0068280320167541505,
153
- 0.006812672138214112,
154
- 0.006838304042816162,
155
- 0.00678604793548584,
156
- 0.0067604160308837894,
157
- 0.006806528091430664,
158
- 0.006788095951080322,
159
- 0.006967296123504638,
160
- 0.00678713607788086,
161
- 0.006849535942077637,
162
- 0.006841343879699707,
163
- 0.006770688056945801,
164
- 0.006865920066833496,
165
- 0.006846464157104492,
166
- 0.0067686400413513184,
167
- 0.006779903888702392,
168
- 0.006781951904296875,
169
- 0.006807551860809326,
170
- 0.0067573761940002445,
171
- 0.006793216228485107,
172
- 0.006767615795135498,
173
- 0.006759456157684326,
174
- 0.006793216228485107,
175
- 0.006770688056945801,
176
- 0.006762495994567871,
177
- 0.006761472225189209,
178
- 0.006776832103729248,
179
- 0.006785024166107178,
180
- 0.006767680168151855,
181
- 0.006807551860809326,
182
- 0.006817791938781738,
183
- 0.00676966381072998,
184
- 0.006779903888702392,
185
- 0.006809599876403808,
186
- 0.006812672138214112,
187
- 0.007303167819976806,
188
- 0.006812672138214112,
189
- 0.0068055038452148435,
190
- 0.006770688056945801,
191
- 0.006782911777496338,
192
- 0.0068055038452148435,
193
- 0.006807551860809326,
194
- 0.006802432060241699,
195
- 0.006810656070709228,
196
- 0.0067870721817016606,
197
- 0.006762495994567871,
198
- 0.006791168212890625,
199
- 0.006796288013458252,
200
- 0.006817791938781738,
201
- 0.006759424209594727,
202
- 0.0067870721817016606,
203
- 0.00677785587310791,
204
- 0.006739967823028564,
205
- 0.006812672138214112,
206
- 0.006806528091430664,
207
- 0.0067573761940002445,
208
- 0.0067717118263244626,
209
- 0.0067983360290527345,
210
- 0.006776832103729248,
211
- 0.00673689603805542,
212
- 0.006763519763946534,
213
- 0.006789120197296142,
214
- 0.006751232147216797,
215
- 0.006800384044647217,
216
- 0.0067645440101623535,
217
- 0.006758399963378906,
218
- 0.006780928134918213,
219
- 0.006800352096557617,
220
- 0.006767615795135498,
221
- 0.006751232147216797,
222
- 0.006737919807434082,
223
- 0.006740992069244385,
224
- 0.006727615833282471,
225
- 0.006754303932189941,
226
- 0.006719488143920899,
227
- 0.006725632190704346,
228
- 0.0068055038452148435,
229
- 0.0067358717918396,
230
- 0.006703104019165039,
231
- 0.006749184131622315,
232
- 0.0067348480224609375,
233
- 0.006729728221893311,
234
- 0.006715392112731934,
235
- 0.0067358717918396,
236
- 0.0067358717918396,
237
- 0.0067276802062988285,
238
- 0.006739967823028564,
239
- 0.006750207901000976,
240
- 0.006717440128326416,
241
- 0.00679423999786377,
242
- 0.006744063854217529,
243
- 0.0067348480224609375,
244
- 0.006790143966674805,
245
- 0.006781951904296875,
246
- 0.0067717118263244626,
247
- 0.006740992069244385,
248
- 0.006745120048522949,
249
- 0.006756351947784424,
250
- 0.006747136116027832,
251
- 0.006763519763946534,
252
- 0.006793216228485107,
253
- 0.006718463897705078,
254
- 0.006751232147216797,
255
- 0.0067276802062988285,
256
- 0.006703104019165039,
257
- 0.006717472076416015,
258
- 0.007165952205657959,
259
- 0.006743040084838867,
260
- 0.006721536159515381,
261
- 0.006823935985565186,
262
- 0.006763519763946534,
263
- 0.006749184131622315,
264
- 0.006754335880279541,
265
- 0.006740992069244385,
266
- 0.006762495994567871,
267
- 0.006765567779541016
268
  ]
269
  },
270
  "throughput": {
271
  "unit": "samples/s",
272
- "value": 146.3975310991604
273
  },
274
  "energy": {
275
  "unit": "kWh",
276
- "cpu": 8.417799000292899e-08,
277
- "ram": 4.6017656379572535e-08,
278
- "gpu": 1.5290201357446433e-07,
279
- "total": 2.830976599569658e-07
280
  },
281
  "efficiency": {
282
  "unit": "samples/kWh",
283
- "value": 3532349.932359073
284
  }
285
  }
286
  }
 
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
 
9
  "model": "FacebookAI/roberta-base",
10
+ "library": "transformers",
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": false
21
+ },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.289728,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
 
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.40.2",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.30.1",
92
  "accelerate_commit": null,
93
  "diffusers_version": "0.27.2",
94
  "diffusers_commit": null,
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 891.00288,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 136,
116
+ "total": 0.9980350713729856,
117
+ "mean": 0.007338493171860189,
118
+ "stdev": 0.00022279241863946134,
119
+ "p50": 0.007393791913986206,
120
+ "p90": 0.007586303949356079,
121
+ "p95": 0.007633935928344727,
122
+ "p99": 0.007742544031143188,
123
  "values": [
124
+ 0.0077619199752807615,
125
+ 0.007643136024475097,
126
+ 0.007706560134887695,
127
+ 0.007701504230499268,
128
+ 0.007520256042480469,
129
+ 0.007599103927612305,
130
+ 0.007539711952209473,
131
+ 0.0075346240997314455,
132
+ 0.007480319976806641,
133
+ 0.007457791805267334,
134
+ 0.007255040168762207,
135
+ 0.007449600219726562,
136
+ 0.007479296207427978,
137
  0.0074291200637817386,
138
+ 0.007436287879943848,
139
+ 0.007308288097381592,
140
+ 0.007264256000518799,
141
+ 0.00733900785446167,
142
+ 0.007513088226318359,
143
+ 0.007528448104858398,
144
+ 0.007599103927612305,
145
+ 0.007576576232910156,
146
+ 0.007632895946502686,
147
+ 0.0075345921516418455,
148
+ 0.007529471874237061,
149
+ 0.007567296028137207,
150
+ 0.0075673599243164065,
151
+ 0.007555071830749512,
152
+ 0.00744755220413208,
153
+ 0.007550975799560547,
154
+ 0.007477248191833496,
155
+ 0.007477248191833496,
156
+ 0.00738918399810791,
157
+ 0.007475168228149414,
158
+ 0.00744652795791626,
159
+ 0.007451648235321045,
160
+ 0.0073994240760803225,
161
+ 0.007502848148345947,
162
+ 0.007490560054779053,
163
+ 0.007546879768371582,
164
+ 0.007637055873870849,
165
+ 0.0075345921516418455,
166
+ 0.007603104114532471,
167
+ 0.007620607852935791,
168
+ 0.0074700798988342285,
169
+ 0.007417856216430664,
170
+ 0.007547904014587403,
171
+ 0.007455743789672851,
172
+ 0.007592959880828858,
173
+ 0.007436287879943848,
174
+ 0.007459839820861816,
175
+ 0.007412735939025879,
176
+ 0.007352255821228028,
177
+ 0.007461887836456299,
178
+ 0.007579648017883301,
179
+ 0.007544832229614258,
180
+ 0.007398399829864502,
181
+ 0.007478271961212158,
182
+ 0.007431168079376221,
183
+ 0.007482367992401123,
184
+ 0.007477248191833496,
185
+ 0.007559167861938477,
186
+ 0.007491583824157715,
187
+ 0.007704576015472412,
188
+ 0.007565311908721924,
189
+ 0.00820633602142334,
190
+ 0.007569407939910889,
191
+ 0.007479296207427978,
192
+ 0.007508992195129394,
193
+ 0.00738099193572998,
194
+ 0.007427072048187256,
195
+ 0.007171072006225586,
196
+ 0.007388160228729248,
197
+ 0.0073134078979492185,
198
+ 0.007341055870056152,
199
+ 0.007418879985809326,
200
+ 0.007269375801086426,
201
+ 0.007488512039184571,
202
+ 0.007251904010772705,
203
+ 0.007277567863464355,
204
+ 0.007612415790557861,
205
+ 0.007341055870056152,
206
+ 0.007350272178649903,
207
+ 0.00748748779296875,
208
+ 0.007237631797790528,
209
+ 0.007096320152282715,
210
+ 0.007063551902770996,
211
+ 0.007126016139984131,
212
+ 0.007218175888061523,
213
+ 0.0070553598403930665,
214
+ 0.007085055828094483,
215
+ 0.007071743965148926,
216
+ 0.007302144050598145,
217
+ 0.007079936027526855,
218
+ 0.007074816226959229,
219
+ 0.007081984043121338,
220
+ 0.007116799831390381,
221
+ 0.007029759883880615,
222
+ 0.007046144008636474,
223
+ 0.007093247890472412,
224
+ 0.007060480117797851,
225
+ 0.0071188478469848635,
226
+ 0.007062528133392334,
227
+ 0.007081984043121338,
228
+ 0.007044095993041993,
229
+ 0.007096320152282715,
230
+ 0.0071792640686035155,
231
+ 0.007088128089904785,
232
+ 0.007048192024230957,
233
+ 0.007065599918365479,
234
+ 0.007117824077606201,
235
+ 0.007157760143280029,
236
+ 0.00703385591506958,
237
+ 0.007079936027526855,
238
+ 0.007080959796905518,
239
+ 0.007085055828094483,
240
+ 0.007066559791564941,
241
+ 0.007065599918365479,
242
+ 0.00709939193725586,
243
+ 0.0071157760620117185,
244
+ 0.007061503887176514,
245
+ 0.007087103843688965,
246
+ 0.007127007961273193,
247
+ 0.007060480117797851,
248
+ 0.007090176105499267,
249
+ 0.007067647933959961,
250
+ 0.007061503887176514,
251
+ 0.007065599918365479,
252
+ 0.007090176105499267,
253
+ 0.007095295906066895,
254
+ 0.007104512214660645,
255
+ 0.0070594558715820314,
256
+ 0.007085055828094483,
257
+ 0.007069695949554444,
258
+ 0.00703385591506958,
259
+ 0.007185408115386963
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
260
  ]
261
  },
262
  "throughput": {
263
  "unit": "samples/s",
264
+ "value": 136.26775641551987
265
  },
266
  "energy": {
267
  "unit": "kWh",
268
+ "cpu": 8.46603555036775e-08,
269
+ "ram": 4.617796958749095e-08,
270
+ "gpu": 1.6257105204255317e-07,
271
+ "total": 2.934093771337216e-07
272
  },
273
  "efficiency": {
274
  "unit": "samples/kWh",
275
+ "value": 3408207.364634597
276
  }
277
  }
278
  }