IlyasMoutawwakil HF staff commited on
Commit
447eff3
·
verified ·
1 Parent(s): b7ac18e

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 750.501888,
108
  "max_global_vram": 1403.518976,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 773.849088,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.39849951171875,
117
- "mean": 7.39849951171875,
118
  "stdev": 0.0,
119
- "p50": 7.39849951171875,
120
- "p90": 7.39849951171875,
121
- "p95": 7.39849951171875,
122
- "p99": 7.39849951171875,
123
  "values": [
124
- 7.39849951171875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 1.1560554506937297e-06,
131
- "ram": 5.914920905689819e-07,
132
- "gpu": 1.6338901959993962e-06,
133
- "total": 3.381437737262108e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1033.539584,
141
  "max_global_vram": 1434.976256,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 794.820608,
@@ -145,105 +145,101 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 74,
149
- "total": 1.0113340473175048,
150
- "mean": 0.013666676315101418,
151
- "stdev": 0.0003348392596776712,
152
- "p50": 0.013616640090942383,
153
- "p90": 0.013874278450012207,
154
- "p95": 0.01418229780197143,
155
- "p99": 0.014985473260879515,
156
  "values": [
157
- 0.014947327613830566,
158
- 0.01467084789276123,
159
- 0.014739456176757813,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
160
  0.013422592163085938,
161
- 0.013218815803527833,
162
- 0.013095935821533204,
163
- 0.013108223915100097,
164
- 0.01306726360321045,
165
- 0.013121503829956055,
166
- 0.01306112003326416,
167
- 0.013455360412597657,
168
- 0.013676544189453126,
169
- 0.013634559631347656,
170
- 0.013602815628051757,
171
- 0.013691904067993164,
172
- 0.013595647811889648,
173
- 0.013636608123779297,
174
- 0.01367142391204834,
175
- 0.013690879821777344,
176
- 0.013610976219177246,
177
- 0.013604864120483399,
178
- 0.013616127967834473,
179
- 0.013575136184692383,
180
- 0.01364684772491455,
181
- 0.013594623565673827,
182
- 0.013624320030212403,
183
- 0.01354854393005371,
184
- 0.013815808296203613,
185
- 0.013594623565673827,
186
- 0.013578240394592284,
187
- 0.013577216148376465,
188
- 0.01365401554107666,
189
- 0.013755392074584961,
190
- 0.013623295783996582,
191
- 0.013567999839782715,
192
- 0.013651968002319336,
193
- 0.013587455749511718,
194
- 0.013616127967834473,
195
- 0.013561856269836426,
196
- 0.015088607788085937,
197
- 0.01365503978729248,
198
- 0.013621248245239258,
199
- 0.01358131217956543,
200
- 0.01359769630432129,
201
- 0.013591551780700683,
202
- 0.013596672058105469,
203
- 0.01359769630432129,
204
- 0.013559807777404785,
205
- 0.013581343650817871,
206
- 0.013872127532958984,
207
- 0.013728768348693847,
208
- 0.013875200271606445,
209
- 0.013919232368469238,
210
- 0.013907967567443847,
211
- 0.013675519943237305,
212
- 0.013741056442260742,
213
- 0.013641728401184081,
214
- 0.013596672058105469,
215
- 0.013661184310913087,
216
- 0.013914112091064454,
217
- 0.013726719856262207,
218
- 0.013557760238647461,
219
- 0.013637632369995116,
220
- 0.013600768089294434,
221
- 0.013643775939941406,
222
- 0.013659135818481445,
223
- 0.013617152214050294,
224
- 0.013658111572265624,
225
- 0.013613056182861329,
226
- 0.013600768089294434,
227
- 0.013586432456970214,
228
- 0.013594623565673827,
229
- 0.013674495697021484,
230
- 0.013644800186157227
231
  ]
232
  },
233
  "throughput": {
234
  "unit": "samples/s",
235
- "value": 73.17068005005862
236
  },
237
  "energy": {
238
  "unit": "kWh",
239
- "cpu": 1.6020326182430784e-07,
240
- "ram": 8.758736850434024e-08,
241
- "gpu": 3.4755658435135294e-07,
242
- "total": 5.95347214680001e-07
243
  },
244
  "efficiency": {
245
  "unit": "samples/kWh",
246
- "value": 1679692.0777356785
247
  }
248
  }
249
  }
 
3
  "name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 773.12,
108
  "max_global_vram": 1403.518976,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 773.849088,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.36253564453125,
117
+ "mean": 7.36253564453125,
118
  "stdev": 0.0,
119
+ "p50": 7.36253564453125,
120
+ "p90": 7.36253564453125,
121
+ "p95": 7.36253564453125,
122
+ "p99": 7.36253564453125,
123
  "values": [
124
+ 7.36253564453125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 1.0754594895841605e-06,
131
+ "ram": 5.732178412276174e-07,
132
+ "gpu": 0.0,
133
+ "total": 1.6486773308117779e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1160.84736,
141
  "max_global_vram": 1434.976256,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 794.820608,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 70,
149
+ "total": 1.0097878379821774,
150
+ "mean": 0.014425540542602538,
151
+ "stdev": 0.0007732916924862301,
152
+ "p50": 0.01476095962524414,
153
+ "p90": 0.015105740451812745,
154
+ "p95": 0.015263743877410888,
155
+ "p99": 0.016176445560455326,
156
  "values": [
157
+ 0.01660108757019043,
158
+ 0.015985664367675782,
159
+ 0.014967871665954589,
160
+ 0.014916607856750488,
161
+ 0.014915583610534668,
162
+ 0.015001600265502929,
163
+ 0.014841856002807617,
164
+ 0.015101951599121094,
165
+ 0.015039487838745117,
166
+ 0.01488486385345459,
167
+ 0.015033344268798828,
168
+ 0.01489401626586914,
169
+ 0.014934016227722167,
170
+ 0.014765055656433105,
171
+ 0.014870528221130372,
172
+ 0.014840831756591797,
173
+ 0.014927871704101562,
174
+ 0.014731264114379883,
175
+ 0.014839712142944337,
176
+ 0.015160320281982421,
177
+ 0.015139840126037597,
178
+ 0.014767104148864747,
179
+ 0.014942208290100097,
180
+ 0.015342592239379883,
181
+ 0.015063039779663086,
182
+ 0.014753791809082031,
183
+ 0.015004672050476075,
184
+ 0.014993408203125,
185
+ 0.014756863594055175,
186
+ 0.014690303802490234,
187
+ 0.014809087753295898,
188
+ 0.01528217601776123,
189
+ 0.015241215705871582,
190
+ 0.015047679901123047,
191
+ 0.014864383697509765,
192
+ 0.014880767822265625,
193
+ 0.014708800315856934,
194
+ 0.014651391983032227,
195
+ 0.014767135620117187,
196
+ 0.014462976455688477,
197
+ 0.014701567649841308,
198
+ 0.014755840301513673,
199
+ 0.01488588809967041,
200
+ 0.014856191635131836,
201
+ 0.014531583786010742,
202
+ 0.013788160324096679,
203
+ 0.013472767829895019,
204
+ 0.013526016235351563,
205
+ 0.013396991729736327,
206
+ 0.013373439788818359,
207
+ 0.013476863861083984,
208
+ 0.013364192008972167,
209
+ 0.013415424346923829,
210
+ 0.013452287673950195,
211
+ 0.01347481632232666,
212
+ 0.013492223739624023,
213
+ 0.013468671798706054,
214
+ 0.013442048072814941,
215
+ 0.013475839614868163,
216
+ 0.013365280151367188,
217
+ 0.01330073642730713,
218
+ 0.013347840309143067,
219
+ 0.013334527969360351,
220
+ 0.013332480430603028,
221
+ 0.013339648246765137,
222
+ 0.013362175941467285,
223
  0.013422592163085938,
224
+ 0.013507583618164062,
225
+ 0.013610943794250489,
226
+ 0.014090239524841308
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
227
  ]
228
  },
229
  "throughput": {
230
  "unit": "samples/s",
231
+ "value": 69.32149246308853
232
  },
233
  "energy": {
234
  "unit": "kWh",
235
+ "cpu": 1.665769153838689e-07,
236
+ "ram": 9.092307769577581e-08,
237
+ "gpu": 3.7156665527778475e-07,
238
+ "total": 6.290666483574295e-07
239
  },
240
  "efficiency": {
241
  "unit": "samples/kWh",
242
+ "value": 1589656.6804346142
243
  }
244
  }
245
  }