IlyasMoutawwakil HF staff commited on
Commit
fa1bed1
1 Parent(s): 432351c

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
- "optimum_benchmark_commit": "dd02f26cb819965cbf86e16d9ce013cddc3b86af",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 937.426944,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,61 +107,57 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 30,
111
- "total": 1.024279958999898,
112
- "mean": 0.03414266529999661,
113
- "stdev": 0.002003813720873649,
114
- "p50": 0.034858600000006845,
115
- "p90": 0.03565418159996057,
116
- "p95": 0.035895198800008646,
117
- "p99": 0.03596436321000283,
118
  "values": [
119
- 0.035967481000000134,
120
- 0.03514002799994387,
121
- 0.035576557000013054,
122
- 0.03432157199995345,
123
- 0.03563575799995533,
124
- 0.03488703400000759,
125
- 0.03473305500000379,
126
- 0.03504126299998234,
127
- 0.03468897199996945,
128
- 0.034071383000025435,
129
- 0.03581999400000768,
130
- 0.03483442499998546,
131
- 0.03480534999999918,
132
- 0.03475111799997421,
133
- 0.03510695500000338,
134
- 0.03486445100003266,
135
- 0.03482678099999248,
136
- 0.035538115000008474,
137
- 0.035070797999992465,
138
- 0.03485274899998103,
139
- 0.035258930999987115,
140
- 0.0355690829999844,
141
- 0.03595673000000943,
142
- 0.03492708800001765,
143
- 0.03258657999998604,
144
- 0.030260984000051394,
145
- 0.029760967000015626,
146
- 0.029760896000027515,
147
- 0.029487383000002865,
148
- 0.030177477999984603
149
  ]
150
  },
151
  "throughput": {
152
  "unit": "samples/s",
153
- "value": 29.288867497995223
154
  },
155
  "energy": {
156
  "unit": "kWh",
157
- "cpu": 1.1727454226001415e-06,
158
- "ram": 4.900842445615982e-08,
159
  "gpu": 0.0,
160
- "total": 1.2217538470563013e-06
161
  },
162
  "efficiency": {
163
  "unit": "samples/kWh",
164
- "value": 818495.47878192
165
  }
166
  }
167
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
+ "optimum_benchmark_commit": "f85421c91b2a2b1bc0a30ceba32f9580499bf023",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 936.914944,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 26,
111
+ "total": 1.0144855889998325,
112
+ "mean": 0.03901867649999356,
113
+ "stdev": 0.0023170659309138014,
114
+ "p50": 0.039327529500013725,
115
+ "p90": 0.04128787400000533,
116
+ "p95": 0.04169794099996693,
117
+ "p99": 0.042700211249950826,
118
  "values": [
119
+ 0.0398925659999918,
120
+ 0.038428609999982655,
121
+ 0.04026809800001274,
122
+ 0.03919330799999443,
123
+ 0.04182835499995008,
124
+ 0.03935162499999478,
125
+ 0.042990829999951075,
126
+ 0.03955073800000264,
127
+ 0.03896483199997647,
128
+ 0.04074107200000299,
129
+ 0.041269048999993174,
130
+ 0.039303434000032667,
131
+ 0.04085590699997965,
132
+ 0.04130669900001749,
133
+ 0.04013485899997704,
134
+ 0.03952430700002196,
135
+ 0.039185905000010735,
136
+ 0.03926400000000285,
137
+ 0.038783142999989195,
138
+ 0.04017673799995691,
139
+ 0.039138707000006434,
140
+ 0.03713710400000991,
141
+ 0.03449093000000403,
142
+ 0.03405434299997978,
143
+ 0.03474384200001168,
144
+ 0.033906587999979365
 
 
 
 
145
  ]
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
+ "value": 25.628752425781663
150
  },
151
  "energy": {
152
  "unit": "kWh",
153
+ "cpu": 1.4029383659362792e-06,
154
+ "ram": 5.86299957871412e-08,
155
  "gpu": 0.0,
156
+ "total": 1.4615683617234204e-06
157
  },
158
  "efficiency": {
159
  "unit": "samples/kWh",
160
+ "value": 684196.5290086342
161
  }
162
  }
163
  }