IlyasMoutawwakil HF staff commited on
Commit
e08a645
1 Parent(s): 0e4db24

Upload cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
@@ -45,7 +45,6 @@
45
  "warmup_runs": 1,
46
  "input_shapes": {
47
  "batch_size": 1,
48
- "num_choices": 2,
49
  "sequence_length": 2
50
  },
51
  "new_tokens": null,
@@ -73,23 +72,23 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
- "cpu_ram_mb": 16766.783488,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-6.8.0-1015-azure-x86_64-with-glibc2.39",
80
  "processor": "x86_64",
81
  "python_version": "3.10.15",
82
  "optimum_benchmark_version": "0.5.0.dev0",
83
- "optimum_benchmark_commit": "08c9f59440cf4e5a5d6711ec19e8329ab2de652d",
84
- "transformers_version": "4.45.2",
85
  "transformers_commit": null,
86
- "accelerate_version": "1.0.1",
87
  "accelerate_commit": null,
88
- "diffusers_version": "0.30.3",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
- "timm_version": "1.0.9",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
@@ -101,7 +100,7 @@
101
  "load": {
102
  "memory": {
103
  "unit": "MB",
104
- "max_ram": 1165.996032,
105
  "max_global_vram": null,
106
  "max_process_vram": null,
107
  "max_reserved": null,
@@ -110,15 +109,15 @@
110
  "latency": {
111
  "unit": "s",
112
  "values": [
113
- 1.7677931680000256
114
  ],
115
  "count": 1,
116
- "total": 1.7677931680000256,
117
- "mean": 1.7677931680000256,
118
- "p50": 1.7677931680000256,
119
- "p90": 1.7677931680000256,
120
- "p95": 1.7677931680000256,
121
- "p99": 1.7677931680000256,
122
  "stdev": 0,
123
  "stdev_": 0
124
  },
@@ -129,7 +128,7 @@
129
  "forward": {
130
  "memory": {
131
  "unit": "MB",
132
- "max_ram": 1189.748736,
133
  "max_global_vram": null,
134
  "max_process_vram": null,
135
  "max_reserved": null,
@@ -138,37 +137,37 @@
138
  "latency": {
139
  "unit": "s",
140
  "values": [
141
- 0.22954940099998566,
142
- 0.22899878599997692,
143
- 0.15467115800004194,
144
- 0.14380051700004515,
145
- 0.142212212000004,
146
- 0.1485327620000021
147
  ],
148
  "count": 6,
149
- "total": 1.0477648360000558,
150
- "mean": 0.17462747266667597,
151
- "p50": 0.15160196000002202,
152
- "p90": 0.2292740934999813,
153
- "p95": 0.22941174724998348,
154
- "p99": 0.22952187024998522,
155
- "stdev": 0.038843482218275785,
156
- "stdev_": 22.243626174685087
157
  },
158
  "throughput": {
159
  "unit": "samples/s",
160
- "value": 5.726475821526502
161
  },
162
  "energy": {
163
  "unit": "kWh",
164
- "cpu": 5.623438200000111e-06,
165
- "ram": 2.3514651591047034e-07,
166
  "gpu": 0.0,
167
- "total": 5.858584715910581e-06
168
  },
169
  "efficiency": {
170
  "unit": "samples/kWh",
171
- "value": 170689.68846421695
172
  }
173
  }
174
  }
 
3
  "name": "cpu_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.5.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
 
45
  "warmup_runs": 1,
46
  "input_shapes": {
47
  "batch_size": 1,
 
48
  "sequence_length": 2
49
  },
50
  "new_tokens": null,
 
72
  "environment": {
73
  "cpu": " AMD EPYC 7763 64-Core Processor",
74
  "cpu_count": 4,
75
+ "cpu_ram_mb": 16757.342208,
76
  "system": "Linux",
77
  "machine": "x86_64",
78
+ "platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
79
  "processor": "x86_64",
80
  "python_version": "3.10.15",
81
  "optimum_benchmark_version": "0.5.0.dev0",
82
+ "optimum_benchmark_commit": "6807ba28334c8c98abf72a03d78f86133328d180",
83
+ "transformers_version": "4.46.3",
84
  "transformers_commit": null,
85
+ "accelerate_version": "1.1.1",
86
  "accelerate_commit": null,
87
+ "diffusers_version": "0.31.0",
88
  "diffusers_commit": null,
89
  "optimum_version": null,
90
  "optimum_commit": null,
91
+ "timm_version": "1.0.11",
92
  "timm_commit": null,
93
  "peft_version": null,
94
  "peft_commit": null
 
100
  "load": {
101
  "memory": {
102
  "unit": "MB",
103
+ "max_ram": 1183.219712,
104
  "max_global_vram": null,
105
  "max_process_vram": null,
106
  "max_reserved": null,
 
109
  "latency": {
110
  "unit": "s",
111
  "values": [
112
+ 1.7816587899999945
113
  ],
114
  "count": 1,
115
+ "total": 1.7816587899999945,
116
+ "mean": 1.7816587899999945,
117
+ "p50": 1.7816587899999945,
118
+ "p90": 1.7816587899999945,
119
+ "p95": 1.7816587899999945,
120
+ "p99": 1.7816587899999945,
121
  "stdev": 0,
122
  "stdev_": 0
123
  },
 
128
  "forward": {
129
  "memory": {
130
  "unit": "MB",
131
+ "max_ram": 1212.719104,
132
  "max_global_vram": null,
133
  "max_process_vram": null,
134
  "max_reserved": null,
 
137
  "latency": {
138
  "unit": "s",
139
  "values": [
140
+ 0.23083957300002567,
141
+ 0.23025468199995203,
142
+ 0.19240379799998664,
143
+ 0.14300760600002604,
144
+ 0.14221200200000794,
145
+ 0.142648136000048
146
  ],
147
  "count": 6,
148
+ "total": 1.0813657970000463,
149
+ "mean": 0.18022763283334106,
150
+ "p50": 0.16770570200000634,
151
+ "p90": 0.23054712749998885,
152
+ "p95": 0.23069335025000726,
153
+ "p99": 0.23081032845002197,
154
+ "stdev": 0.039697334715932016,
155
+ "stdev_": 22.02621989306195
156
  },
157
  "throughput": {
158
  "unit": "samples/s",
159
+ "value": 5.5485387244957804
160
  },
161
  "energy": {
162
  "unit": "kWh",
163
+ "cpu": 5.5791582555558406e-06,
164
+ "ram": 2.3316236891095142e-07,
165
  "gpu": 0.0,
166
+ "total": 5.812320624466793e-06
167
  },
168
  "efficiency": {
169
  "unit": "samples/kWh",
170
+ "value": 172048.32021663248
171
  }
172
  }
173
  }