IlyasMoutawwakil HF staff commited on
Commit
4ac748f
1 Parent(s): 4415744

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/file.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/file.json ADDED
@@ -0,0 +1,297 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.2.2",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-classification",
9
+ "model": "FacebookAI/roberta-base",
10
+ "library": "transformers",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": false
21
+ },
22
+ "no_weights": true,
23
+ "device_map": null,
24
+ "torch_dtype": null,
25
+ "eval_mode": true,
26
+ "to_bettertransformer": false,
27
+ "low_cpu_mem_usage": null,
28
+ "attn_implementation": null,
29
+ "cache_implementation": null,
30
+ "autocast_enabled": false,
31
+ "autocast_dtype": null,
32
+ "torch_compile": false,
33
+ "torch_compile_target": "forward",
34
+ "torch_compile_config": {},
35
+ "quantization_scheme": null,
36
+ "quantization_config": {},
37
+ "deepspeed_inference": false,
38
+ "deepspeed_inference_config": {},
39
+ "peft_type": null,
40
+ "peft_config": {}
41
+ },
42
+ "scenario": {
43
+ "name": "inference",
44
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
45
+ "iterations": 1,
46
+ "duration": 1,
47
+ "warmup_runs": 1,
48
+ "input_shapes": {
49
+ "batch_size": 1,
50
+ "num_choices": 2,
51
+ "sequence_length": 2
52
+ },
53
+ "new_tokens": null,
54
+ "latency": true,
55
+ "memory": true,
56
+ "energy": true,
57
+ "forward_kwargs": {},
58
+ "generate_kwargs": {
59
+ "max_new_tokens": 2,
60
+ "min_new_tokens": 2
61
+ },
62
+ "call_kwargs": {
63
+ "num_inference_steps": 2
64
+ }
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "error",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
80
+ "processor": "x86_64",
81
+ "python_version": "3.10.14",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.0",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.40.2",
90
+ "transformers_commit": null,
91
+ "accelerate_version": "0.30.0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": "0.27.2",
94
+ "diffusers_commit": null,
95
+ "optimum_version": null,
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": null,
100
+ "peft_commit": null
101
+ }
102
+ },
103
+ "report": {
104
+ "forward": {
105
+ "memory": {
106
+ "unit": "MB",
107
+ "max_ram": 890.138624,
108
+ "max_global_vram": 1195.900928,
109
+ "max_process_vram": 0.0,
110
+ "max_reserved": 555.74528,
111
+ "max_allocated": 508.993536
112
+ },
113
+ "latency": {
114
+ "unit": "s",
115
+ "count": 154,
116
+ "total": 1.0014327363967899,
117
+ "mean": 0.00650280997660253,
118
+ "stdev": 0.00021388913393038193,
119
+ "p50": 0.006459903955459595,
120
+ "p90": 0.006831820774078369,
121
+ "p95": 0.006857471919059753,
122
+ "p99": 0.0069436825037002565,
123
+ "values": [
124
+ 0.007262207984924316,
125
+ 0.0067983360290527345,
126
+ 0.006879231929779053,
127
+ 0.006853631973266602,
128
+ 0.006860799789428711,
129
+ 0.006831103801727295,
130
+ 0.00682700777053833,
131
+ 0.00685152006149292,
132
+ 0.00694271993637085,
133
+ 0.006814720153808594,
134
+ 0.006944767951965332,
135
+ 0.006825984001159668,
136
+ 0.006788095951080322,
137
+ 0.006792191982269287,
138
+ 0.0067041277885437015,
139
+ 0.006729728221893311,
140
+ 0.006793216228485107,
141
+ 0.006850560188293457,
142
+ 0.006896639823913574,
143
+ 0.006820864200592041,
144
+ 0.006834176063537598,
145
+ 0.006816768169403077,
146
+ 0.006791168212890625,
147
+ 0.0068321280479431154,
148
+ 0.006855679988861084,
149
+ 0.006842368125915528,
150
+ 0.006790143966674805,
151
+ 0.00682700777053833,
152
+ 0.006816768169403077,
153
+ 0.006427648067474365,
154
+ 0.006453248023986816,
155
+ 0.006552576065063476,
156
+ 0.006458367824554443,
157
+ 0.006496223926544189,
158
+ 0.006458367824554443,
159
+ 0.00657203197479248,
160
+ 0.006557695865631104,
161
+ 0.006412288188934326,
162
+ 0.0064204797744750975,
163
+ 0.0063907837867736815,
164
+ 0.006481919765472412,
165
+ 0.006502431869506836,
166
+ 0.006575104236602783,
167
+ 0.006882304191589355,
168
+ 0.0068618240356445315,
169
+ 0.006852608203887939,
170
+ 0.006523903846740723,
171
+ 0.0064767999649047855,
172
+ 0.006418432235717773,
173
+ 0.0064440321922302245,
174
+ 0.006808576107025147,
175
+ 0.006523903846740723,
176
+ 0.006467584133148193,
177
+ 0.006665184020996094,
178
+ 0.006419456005096436,
179
+ 0.006407167911529541,
180
+ 0.006486015796661377,
181
+ 0.006436863899230957,
182
+ 0.006506559848785401,
183
+ 0.0065781760215759275,
184
+ 0.006404096126556396,
185
+ 0.006455296039581298,
186
+ 0.006433792114257812,
187
+ 0.006498303890228272,
188
+ 0.006457344055175781,
189
+ 0.006512639999389648,
190
+ 0.006477759838104248,
191
+ 0.006490111827850342,
192
+ 0.006460415840148926,
193
+ 0.00658739185333252,
194
+ 0.006543360233306885,
195
+ 0.006508480072021485,
196
+ 0.006599679946899414,
197
+ 0.006478847980499268,
198
+ 0.006461440086364746,
199
+ 0.006533120155334473,
200
+ 0.006404096126556396,
201
+ 0.006509568214416504,
202
+ 0.006450175762176514,
203
+ 0.006375423908233643,
204
+ 0.00636518383026123,
205
+ 0.006194176197052002,
206
+ 0.006404096126556396,
207
+ 0.006467584133148193,
208
+ 0.006201312065124511,
209
+ 0.006201344013214111,
210
+ 0.006171648025512695,
211
+ 0.006261760234832763,
212
+ 0.006228991985321045,
213
+ 0.006301695823669433,
214
+ 0.006254591941833496,
215
+ 0.0063303041458129886,
216
+ 0.006481919765472412,
217
+ 0.006253568172454834,
218
+ 0.00618393611907959,
219
+ 0.006127615928649902,
220
+ 0.006198272228240966,
221
+ 0.006131711959838867,
222
+ 0.006128640174865723,
223
+ 0.006179840087890625,
224
+ 0.006153215885162353,
225
+ 0.00618393611907959,
226
+ 0.006137792110443115,
227
+ 0.006278143882751465,
228
+ 0.00611737585067749,
229
+ 0.006278143882751465,
230
+ 0.006145023822784424,
231
+ 0.00614192008972168,
232
+ 0.0062494721412658695,
233
+ 0.006147039890289307,
234
+ 0.0066826238632202144,
235
+ 0.006419456005096436,
236
+ 0.006433792114257812,
237
+ 0.006445055961608887,
238
+ 0.006451200008392334,
239
+ 0.00652288007736206,
240
+ 0.006462463855743408,
241
+ 0.006412288188934326,
242
+ 0.00643071985244751,
243
+ 0.006379519939422608,
244
+ 0.00643174409866333,
245
+ 0.0064174079895019534,
246
+ 0.00638976001739502,
247
+ 0.006446080207824707,
248
+ 0.00637440013885498,
249
+ 0.006553599834442139,
250
+ 0.006446080207824707,
251
+ 0.006403071880340576,
252
+ 0.006481919765472412,
253
+ 0.00643174409866333,
254
+ 0.006437888145446777,
255
+ 0.006496255874633789,
256
+ 0.006459392070770263,
257
+ 0.00648089599609375,
258
+ 0.006452223777770996,
259
+ 0.006419424057006836,
260
+ 0.006425663948059082,
261
+ 0.006435840129852295,
262
+ 0.006437888145446777,
263
+ 0.006432767868041992,
264
+ 0.006435840129852295,
265
+ 0.006478847980499268,
266
+ 0.0064287037849426265,
267
+ 0.0064430079460144046,
268
+ 0.0064778242111206055,
269
+ 0.006423552036285401,
270
+ 0.006436863899230957,
271
+ 0.006426623821258545,
272
+ 0.0064471039772033695,
273
+ 0.006494175910949707,
274
+ 0.00648089599609375,
275
+ 0.006475776195526123,
276
+ 0.006486015796661377,
277
+ 0.006469632148742676
278
+ ]
279
+ },
280
+ "throughput": {
281
+ "unit": "samples/s",
282
+ "value": 153.7796742635961
283
+ },
284
+ "energy": {
285
+ "unit": "kWh",
286
+ "cpu": 7.514877954783488e-08,
287
+ "ram": 4.1082792785910214e-08,
288
+ "gpu": 1.3599166997468262e-07,
289
+ "total": 2.5222324230842775e-07
290
+ },
291
+ "efficiency": {
292
+ "unit": "samples/kWh",
293
+ "value": 3964741.674271095
294
+ }
295
+ }
296
+ }
297
+ }