Upload cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json with huggingface_hub
Browse files
cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
-
"version": "2.4.
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "text-generation",
|
9 |
"library": "transformers",
|
@@ -110,7 +110,7 @@
|
|
110 |
"overall": {
|
111 |
"memory": {
|
112 |
"unit": "MB",
|
113 |
-
"max_ram": 1330.
|
114 |
"max_global_vram": 3559.391232,
|
115 |
"max_process_vram": 0.0,
|
116 |
"max_reserved": 2910.846976,
|
@@ -119,24 +119,24 @@
|
|
119 |
"latency": {
|
120 |
"unit": "s",
|
121 |
"count": 5,
|
122 |
-
"total": 0.
|
123 |
-
"mean": 0.
|
124 |
-
"stdev": 0.
|
125 |
-
"p50": 0.
|
126 |
-
"p90": 0.
|
127 |
-
"p95": 0.
|
128 |
-
"p99": 0.
|
129 |
"values": [
|
130 |
-
0.
|
131 |
-
0.
|
132 |
-
0.
|
133 |
-
0.
|
134 |
-
0.
|
135 |
]
|
136 |
},
|
137 |
"throughput": {
|
138 |
"unit": "samples/s",
|
139 |
-
"value": 94.
|
140 |
},
|
141 |
"energy": null,
|
142 |
"efficiency": null
|
@@ -144,7 +144,7 @@
|
|
144 |
"warmup": {
|
145 |
"memory": {
|
146 |
"unit": "MB",
|
147 |
-
"max_ram": 1330.
|
148 |
"max_global_vram": 3559.391232,
|
149 |
"max_process_vram": 0.0,
|
150 |
"max_reserved": 2910.846976,
|
@@ -153,21 +153,21 @@
|
|
153 |
"latency": {
|
154 |
"unit": "s",
|
155 |
"count": 2,
|
156 |
-
"total": 0.
|
157 |
-
"mean": 0.
|
158 |
-
"stdev": 0.
|
159 |
-
"p50": 0.
|
160 |
-
"p90": 0.
|
161 |
-
"p95": 0.
|
162 |
-
"p99": 0.
|
163 |
"values": [
|
164 |
-
0.
|
165 |
-
0.
|
166 |
]
|
167 |
},
|
168 |
"throughput": {
|
169 |
"unit": "samples/s",
|
170 |
-
"value": 20.
|
171 |
},
|
172 |
"energy": null,
|
173 |
"efficiency": null
|
@@ -175,7 +175,7 @@
|
|
175 |
"train": {
|
176 |
"memory": {
|
177 |
"unit": "MB",
|
178 |
-
"max_ram": 1330.
|
179 |
"max_global_vram": 3559.391232,
|
180 |
"max_process_vram": 0.0,
|
181 |
"max_reserved": 2910.846976,
|
@@ -184,22 +184,22 @@
|
|
184 |
"latency": {
|
185 |
"unit": "s",
|
186 |
"count": 3,
|
187 |
-
"total": 0.
|
188 |
-
"mean": 0.
|
189 |
-
"stdev":
|
190 |
-
"p50": 0.
|
191 |
-
"p90": 0.
|
192 |
-
"p95": 0.
|
193 |
-
"p99": 0.
|
194 |
"values": [
|
195 |
-
0.
|
196 |
-
0.
|
197 |
-
0.
|
198 |
]
|
199 |
},
|
200 |
"throughput": {
|
201 |
"unit": "samples/s",
|
202 |
-
"value": 137.
|
203 |
},
|
204 |
"energy": null,
|
205 |
"efficiency": null
|
|
|
3 |
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
+
"version": "2.4.1+cu124",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "text-generation",
|
9 |
"library": "transformers",
|
|
|
110 |
"overall": {
|
111 |
"memory": {
|
112 |
"unit": "MB",
|
113 |
+
"max_ram": 1330.839552,
|
114 |
"max_global_vram": 3559.391232,
|
115 |
"max_process_vram": 0.0,
|
116 |
"max_reserved": 2910.846976,
|
|
|
119 |
"latency": {
|
120 |
"unit": "s",
|
121 |
"count": 5,
|
122 |
+
"total": 0.5287597389221191,
|
123 |
+
"mean": 0.10575194778442383,
|
124 |
+
"stdev": 0.1238597264813096,
|
125 |
+
"p50": 0.04361312103271484,
|
126 |
+
"p90": 0.22987570953369144,
|
127 |
+
"p95": 0.29167308425903316,
|
128 |
+
"p99": 0.3411109840393066,
|
129 |
"values": [
|
130 |
+
0.353470458984375,
|
131 |
+
0.044483585357666014,
|
132 |
+
0.04361312103271484,
|
133 |
+
0.04359372711181641,
|
134 |
+
0.04359884643554687
|
135 |
]
|
136 |
},
|
137 |
"throughput": {
|
138 |
"unit": "samples/s",
|
139 |
+
"value": 94.5609060590078
|
140 |
},
|
141 |
"energy": null,
|
142 |
"efficiency": null
|
|
|
144 |
"warmup": {
|
145 |
"memory": {
|
146 |
"unit": "MB",
|
147 |
+
"max_ram": 1330.839552,
|
148 |
"max_global_vram": 3559.391232,
|
149 |
"max_process_vram": 0.0,
|
150 |
"max_reserved": 2910.846976,
|
|
|
153 |
"latency": {
|
154 |
"unit": "s",
|
155 |
"count": 2,
|
156 |
+
"total": 0.397954044342041,
|
157 |
+
"mean": 0.1989770221710205,
|
158 |
+
"stdev": 0.15449343681335448,
|
159 |
+
"p50": 0.1989770221710205,
|
160 |
+
"p90": 0.32257177162170414,
|
161 |
+
"p95": 0.3380211153030396,
|
162 |
+
"p99": 0.3503805902481079,
|
163 |
"values": [
|
164 |
+
0.353470458984375,
|
165 |
+
0.044483585357666014
|
166 |
]
|
167 |
},
|
168 |
"throughput": {
|
169 |
"unit": "samples/s",
|
170 |
+
"value": 20.10282371480062
|
171 |
},
|
172 |
"energy": null,
|
173 |
"efficiency": null
|
|
|
175 |
"train": {
|
176 |
"memory": {
|
177 |
"unit": "MB",
|
178 |
+
"max_ram": 1330.839552,
|
179 |
"max_global_vram": 3559.391232,
|
180 |
"max_process_vram": 0.0,
|
181 |
"max_reserved": 2910.846976,
|
|
|
184 |
"latency": {
|
185 |
"unit": "s",
|
186 |
"count": 3,
|
187 |
+
"total": 0.13080569458007812,
|
188 |
+
"mean": 0.04360189819335938,
|
189 |
+
"stdev": 8.206337439779797e-06,
|
190 |
+
"p50": 0.04359884643554687,
|
191 |
+
"p90": 0.04361026611328125,
|
192 |
+
"p95": 0.043611693572998045,
|
193 |
+
"p99": 0.043612835540771486,
|
194 |
"values": [
|
195 |
+
0.04361312103271484,
|
196 |
+
0.04359372711181641,
|
197 |
+
0.04359884643554687
|
198 |
]
|
199 |
},
|
200 |
"throughput": {
|
201 |
"unit": "samples/s",
|
202 |
+
"value": 137.60868789225805
|
203 |
},
|
204 |
"energy": null,
|
205 |
"efficiency": null
|