Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
-
"version": "2.4.0+
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "token-classification",
|
9 |
"library": "transformers",
|
@@ -104,7 +104,7 @@
|
|
104 |
"load": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
-
"max_ram":
|
108 |
"max_global_vram": 1403.518976,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 773.849088,
|
@@ -113,31 +113,31 @@
|
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
"count": 1,
|
116 |
-
"total": 7.
|
117 |
-
"mean": 7.
|
118 |
"stdev": 0.0,
|
119 |
-
"p50": 7.
|
120 |
-
"p90": 7.
|
121 |
-
"p95": 7.
|
122 |
-
"p99": 7.
|
123 |
"values": [
|
124 |
-
7.
|
125 |
]
|
126 |
},
|
127 |
"throughput": null,
|
128 |
"energy": {
|
129 |
"unit": "kWh",
|
130 |
-
"cpu": 1.
|
131 |
-
"ram": 5.
|
132 |
-
"gpu":
|
133 |
-
"total":
|
134 |
},
|
135 |
"efficiency": null
|
136 |
},
|
137 |
"forward": {
|
138 |
"memory": {
|
139 |
"unit": "MB",
|
140 |
-
"max_ram":
|
141 |
"max_global_vram": 1434.976256,
|
142 |
"max_process_vram": 0.0,
|
143 |
"max_reserved": 794.820608,
|
@@ -145,105 +145,101 @@
|
|
145 |
},
|
146 |
"latency": {
|
147 |
"unit": "s",
|
148 |
-
"count":
|
149 |
-
"total": 1.
|
150 |
-
"mean": 0.
|
151 |
-
"stdev": 0.
|
152 |
-
"p50": 0.
|
153 |
-
"p90": 0.
|
154 |
-
"p95": 0.
|
155 |
-
"p99": 0.
|
156 |
"values": [
|
157 |
-
0.
|
158 |
-
0.
|
159 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
160 |
0.013422592163085938,
|
161 |
-
0.
|
162 |
-
0.
|
163 |
-
0.
|
164 |
-
0.01306726360321045,
|
165 |
-
0.013121503829956055,
|
166 |
-
0.01306112003326416,
|
167 |
-
0.013455360412597657,
|
168 |
-
0.013676544189453126,
|
169 |
-
0.013634559631347656,
|
170 |
-
0.013602815628051757,
|
171 |
-
0.013691904067993164,
|
172 |
-
0.013595647811889648,
|
173 |
-
0.013636608123779297,
|
174 |
-
0.01367142391204834,
|
175 |
-
0.013690879821777344,
|
176 |
-
0.013610976219177246,
|
177 |
-
0.013604864120483399,
|
178 |
-
0.013616127967834473,
|
179 |
-
0.013575136184692383,
|
180 |
-
0.01364684772491455,
|
181 |
-
0.013594623565673827,
|
182 |
-
0.013624320030212403,
|
183 |
-
0.01354854393005371,
|
184 |
-
0.013815808296203613,
|
185 |
-
0.013594623565673827,
|
186 |
-
0.013578240394592284,
|
187 |
-
0.013577216148376465,
|
188 |
-
0.01365401554107666,
|
189 |
-
0.013755392074584961,
|
190 |
-
0.013623295783996582,
|
191 |
-
0.013567999839782715,
|
192 |
-
0.013651968002319336,
|
193 |
-
0.013587455749511718,
|
194 |
-
0.013616127967834473,
|
195 |
-
0.013561856269836426,
|
196 |
-
0.015088607788085937,
|
197 |
-
0.01365503978729248,
|
198 |
-
0.013621248245239258,
|
199 |
-
0.01358131217956543,
|
200 |
-
0.01359769630432129,
|
201 |
-
0.013591551780700683,
|
202 |
-
0.013596672058105469,
|
203 |
-
0.01359769630432129,
|
204 |
-
0.013559807777404785,
|
205 |
-
0.013581343650817871,
|
206 |
-
0.013872127532958984,
|
207 |
-
0.013728768348693847,
|
208 |
-
0.013875200271606445,
|
209 |
-
0.013919232368469238,
|
210 |
-
0.013907967567443847,
|
211 |
-
0.013675519943237305,
|
212 |
-
0.013741056442260742,
|
213 |
-
0.013641728401184081,
|
214 |
-
0.013596672058105469,
|
215 |
-
0.013661184310913087,
|
216 |
-
0.013914112091064454,
|
217 |
-
0.013726719856262207,
|
218 |
-
0.013557760238647461,
|
219 |
-
0.013637632369995116,
|
220 |
-
0.013600768089294434,
|
221 |
-
0.013643775939941406,
|
222 |
-
0.013659135818481445,
|
223 |
-
0.013617152214050294,
|
224 |
-
0.013658111572265624,
|
225 |
-
0.013613056182861329,
|
226 |
-
0.013600768089294434,
|
227 |
-
0.013586432456970214,
|
228 |
-
0.013594623565673827,
|
229 |
-
0.013674495697021484,
|
230 |
-
0.013644800186157227
|
231 |
]
|
232 |
},
|
233 |
"throughput": {
|
234 |
"unit": "samples/s",
|
235 |
-
"value":
|
236 |
},
|
237 |
"energy": {
|
238 |
"unit": "kWh",
|
239 |
-
"cpu": 1.
|
240 |
-
"ram":
|
241 |
-
"gpu": 3.
|
242 |
-
"total":
|
243 |
},
|
244 |
"efficiency": {
|
245 |
"unit": "samples/kWh",
|
246 |
-
"value":
|
247 |
}
|
248 |
}
|
249 |
}
|
|
|
3 |
"name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
+
"version": "2.4.0+cu124",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "token-classification",
|
9 |
"library": "transformers",
|
|
|
104 |
"load": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
+
"max_ram": 773.12,
|
108 |
"max_global_vram": 1403.518976,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 773.849088,
|
|
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
"count": 1,
|
116 |
+
"total": 7.36253564453125,
|
117 |
+
"mean": 7.36253564453125,
|
118 |
"stdev": 0.0,
|
119 |
+
"p50": 7.36253564453125,
|
120 |
+
"p90": 7.36253564453125,
|
121 |
+
"p95": 7.36253564453125,
|
122 |
+
"p99": 7.36253564453125,
|
123 |
"values": [
|
124 |
+
7.36253564453125
|
125 |
]
|
126 |
},
|
127 |
"throughput": null,
|
128 |
"energy": {
|
129 |
"unit": "kWh",
|
130 |
+
"cpu": 1.0754594895841605e-06,
|
131 |
+
"ram": 5.732178412276174e-07,
|
132 |
+
"gpu": 0.0,
|
133 |
+
"total": 1.6486773308117779e-06
|
134 |
},
|
135 |
"efficiency": null
|
136 |
},
|
137 |
"forward": {
|
138 |
"memory": {
|
139 |
"unit": "MB",
|
140 |
+
"max_ram": 1160.84736,
|
141 |
"max_global_vram": 1434.976256,
|
142 |
"max_process_vram": 0.0,
|
143 |
"max_reserved": 794.820608,
|
|
|
145 |
},
|
146 |
"latency": {
|
147 |
"unit": "s",
|
148 |
+
"count": 70,
|
149 |
+
"total": 1.0097878379821774,
|
150 |
+
"mean": 0.014425540542602538,
|
151 |
+
"stdev": 0.0007732916924862301,
|
152 |
+
"p50": 0.01476095962524414,
|
153 |
+
"p90": 0.015105740451812745,
|
154 |
+
"p95": 0.015263743877410888,
|
155 |
+
"p99": 0.016176445560455326,
|
156 |
"values": [
|
157 |
+
0.01660108757019043,
|
158 |
+
0.015985664367675782,
|
159 |
+
0.014967871665954589,
|
160 |
+
0.014916607856750488,
|
161 |
+
0.014915583610534668,
|
162 |
+
0.015001600265502929,
|
163 |
+
0.014841856002807617,
|
164 |
+
0.015101951599121094,
|
165 |
+
0.015039487838745117,
|
166 |
+
0.01488486385345459,
|
167 |
+
0.015033344268798828,
|
168 |
+
0.01489401626586914,
|
169 |
+
0.014934016227722167,
|
170 |
+
0.014765055656433105,
|
171 |
+
0.014870528221130372,
|
172 |
+
0.014840831756591797,
|
173 |
+
0.014927871704101562,
|
174 |
+
0.014731264114379883,
|
175 |
+
0.014839712142944337,
|
176 |
+
0.015160320281982421,
|
177 |
+
0.015139840126037597,
|
178 |
+
0.014767104148864747,
|
179 |
+
0.014942208290100097,
|
180 |
+
0.015342592239379883,
|
181 |
+
0.015063039779663086,
|
182 |
+
0.014753791809082031,
|
183 |
+
0.015004672050476075,
|
184 |
+
0.014993408203125,
|
185 |
+
0.014756863594055175,
|
186 |
+
0.014690303802490234,
|
187 |
+
0.014809087753295898,
|
188 |
+
0.01528217601776123,
|
189 |
+
0.015241215705871582,
|
190 |
+
0.015047679901123047,
|
191 |
+
0.014864383697509765,
|
192 |
+
0.014880767822265625,
|
193 |
+
0.014708800315856934,
|
194 |
+
0.014651391983032227,
|
195 |
+
0.014767135620117187,
|
196 |
+
0.014462976455688477,
|
197 |
+
0.014701567649841308,
|
198 |
+
0.014755840301513673,
|
199 |
+
0.01488588809967041,
|
200 |
+
0.014856191635131836,
|
201 |
+
0.014531583786010742,
|
202 |
+
0.013788160324096679,
|
203 |
+
0.013472767829895019,
|
204 |
+
0.013526016235351563,
|
205 |
+
0.013396991729736327,
|
206 |
+
0.013373439788818359,
|
207 |
+
0.013476863861083984,
|
208 |
+
0.013364192008972167,
|
209 |
+
0.013415424346923829,
|
210 |
+
0.013452287673950195,
|
211 |
+
0.01347481632232666,
|
212 |
+
0.013492223739624023,
|
213 |
+
0.013468671798706054,
|
214 |
+
0.013442048072814941,
|
215 |
+
0.013475839614868163,
|
216 |
+
0.013365280151367188,
|
217 |
+
0.01330073642730713,
|
218 |
+
0.013347840309143067,
|
219 |
+
0.013334527969360351,
|
220 |
+
0.013332480430603028,
|
221 |
+
0.013339648246765137,
|
222 |
+
0.013362175941467285,
|
223 |
0.013422592163085938,
|
224 |
+
0.013507583618164062,
|
225 |
+
0.013610943794250489,
|
226 |
+
0.014090239524841308
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
227 |
]
|
228 |
},
|
229 |
"throughput": {
|
230 |
"unit": "samples/s",
|
231 |
+
"value": 69.32149246308853
|
232 |
},
|
233 |
"energy": {
|
234 |
"unit": "kWh",
|
235 |
+
"cpu": 1.665769153838689e-07,
|
236 |
+
"ram": 9.092307769577581e-08,
|
237 |
+
"gpu": 3.7156665527778475e-07,
|
238 |
+
"total": 6.290666483574295e-07
|
239 |
},
|
240 |
"efficiency": {
|
241 |
"unit": "samples/kWh",
|
242 |
+
"value": 1589656.6804346142
|
243 |
}
|
244 |
}
|
245 |
}
|