IlyasMoutawwakil HF staff commited on
Commit
2271b00
·
verified ·
1 Parent(s): 05e3828

Upload cuda_training_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_training_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -76,10 +76,10 @@
76
  "environment": {
77
  "cpu": " AMD EPYC 7R32",
78
  "cpu_count": 16,
79
- "cpu_ram_mb": 66697.29792,
80
  "system": "Linux",
81
  "machine": "x86_64",
82
- "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
  "python_version": "3.10.12",
85
  "gpu": [
@@ -89,15 +89,15 @@
89
  "gpu_vram_mb": 24146608128,
90
  "optimum_benchmark_version": "0.2.1",
91
  "optimum_benchmark_commit": null,
92
- "transformers_version": "4.41.1",
93
  "transformers_commit": null,
94
- "accelerate_version": "0.30.1",
95
  "accelerate_commit": null,
96
- "diffusers_version": "0.27.2",
97
  "diffusers_commit": null,
98
  "optimum_version": null,
99
  "optimum_commit": null,
100
- "timm_version": "1.0.3",
101
  "timm_commit": null,
102
  "peft_version": null,
103
  "peft_commit": null
@@ -107,7 +107,7 @@
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
- "max_ram": 1095.077888,
111
  "max_global_vram": 3379.03616,
112
  "max_process_vram": 0.0,
113
  "max_reserved": 2730.491904,
@@ -116,24 +116,24 @@
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
- "total": 1.094946807861328,
120
- "mean": 0.21898936157226562,
121
- "stdev": 0.28533450851903186,
122
- "p50": 0.07649485015869141,
123
- "p90": 0.5046370269775391,
124
- "p95": 0.6471473098754882,
125
- "p99": 0.7611555361938477,
126
  "values": [
127
- 0.7896575927734375,
128
- 0.07649485015869141,
129
- 0.0771061782836914,
130
- 0.07594290924072265,
131
- 0.07574527740478515
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
- "value": 45.664318705728725
137
  },
138
  "energy": null,
139
  "efficiency": null
@@ -141,7 +141,7 @@
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
- "max_ram": 1095.077888,
145
  "max_global_vram": 3379.03616,
146
  "max_process_vram": 0.0,
147
  "max_reserved": 2730.491904,
@@ -150,21 +150,21 @@
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
- "total": 0.8661524429321289,
154
- "mean": 0.43307622146606445,
155
- "stdev": 0.35658137130737305,
156
- "p50": 0.43307622146606445,
157
- "p90": 0.7183413185119629,
158
- "p95": 0.7539994556427002,
159
- "p99": 0.78252596534729,
160
  "values": [
161
- 0.7896575927734375,
162
- 0.07649485015869141
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
- "value": 9.236249421543079
168
  },
169
  "energy": null,
170
  "efficiency": null
@@ -172,7 +172,7 @@
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
- "max_ram": 1095.077888,
176
  "max_global_vram": 3379.03616,
177
  "max_process_vram": 0.0,
178
  "max_reserved": 2730.491904,
@@ -181,22 +181,22 @@
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
- "total": 0.2287943649291992,
185
- "mean": 0.07626478830973306,
186
- "stdev": 0.000600398424299626,
187
- "p50": 0.07594290924072265,
188
- "p90": 0.07687352447509765,
189
- "p95": 0.07698985137939453,
190
- "p99": 0.07708291290283202,
191
  "values": [
192
- 0.0771061782836914,
193
- 0.07594290924072265,
194
- 0.07574527740478515
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 78.67326629993764
200
  },
201
  "energy": null,
202
  "efficiency": null
 
3
  "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
76
  "environment": {
77
  "cpu": " AMD EPYC 7R32",
78
  "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.293824,
80
  "system": "Linux",
81
  "machine": "x86_64",
82
+ "platform": "Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
  "python_version": "3.10.12",
85
  "gpu": [
 
89
  "gpu_vram_mb": 24146608128,
90
  "optimum_benchmark_version": "0.2.1",
91
  "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.42.3",
93
  "transformers_commit": null,
94
+ "accelerate_version": "0.31.0",
95
  "accelerate_commit": null,
96
+ "diffusers_version": "0.29.2",
97
  "diffusers_commit": null,
98
  "optimum_version": null,
99
  "optimum_commit": null,
100
+ "timm_version": "1.0.7",
101
  "timm_commit": null,
102
  "peft_version": null,
103
  "peft_commit": null
 
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
+ "max_ram": 1722.75712,
111
  "max_global_vram": 3379.03616,
112
  "max_process_vram": 0.0,
113
  "max_reserved": 2730.491904,
 
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
+ "total": 0.9112914161682129,
120
+ "mean": 0.18225828323364257,
121
+ "stdev": 0.2655902629849097,
122
+ "p50": 0.0495810546875,
123
+ "p90": 0.4481149078369141,
124
+ "p95": 0.5807765701293944,
125
+ "p99": 0.6869058999633789,
126
  "values": [
127
+ 0.713438232421875,
128
+ 0.050129920959472656,
129
+ 0.0495810546875,
130
+ 0.04907622528076172,
131
+ 0.04906598281860351
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
+ "value": 54.867190794180196
137
  },
138
  "energy": null,
139
  "efficiency": null
 
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
+ "max_ram": 1722.75712,
145
  "max_global_vram": 3379.03616,
146
  "max_process_vram": 0.0,
147
  "max_reserved": 2730.491904,
 
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
+ "total": 0.7635681533813476,
154
+ "mean": 0.3817840766906738,
155
+ "stdev": 0.33165415573120116,
156
+ "p50": 0.3817840766906738,
157
+ "p90": 0.6471074012756348,
158
+ "p95": 0.6802728168487548,
159
+ "p99": 0.7068051493072509,
160
  "values": [
161
+ 0.713438232421875,
162
+ 0.050129920959472656
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
+ "value": 10.477126324052664
168
  },
169
  "energy": null,
170
  "efficiency": null
 
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
+ "max_ram": 1722.75712,
176
  "max_global_vram": 3379.03616,
177
  "max_process_vram": 0.0,
178
  "max_reserved": 2730.491904,
 
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
+ "total": 0.14772326278686523,
185
+ "mean": 0.04924108759562174,
186
+ "stdev": 0.000240429400140505,
187
+ "p50": 0.04907622528076172,
188
+ "p90": 0.049480088806152345,
189
+ "p95": 0.04953057174682617,
190
+ "p99": 0.04957095809936524,
191
  "values": [
192
+ 0.0495810546875,
193
+ 0.04907622528076172,
194
+ 0.04906598281860351
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
+ "value": 121.8494613537636
200
  },
201
  "energy": null,
202
  "efficiency": null