IlyasMoutawwakil HF staff commited on
Commit
8264ddb
1 Parent(s): 54462ee

Upload cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -7,6 +7,7 @@
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
 
10
  "model": "microsoft/deberta-v3-base",
11
  "processor": "microsoft/deberta-v3-base",
12
  "device": "cpu",
@@ -73,14 +74,14 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
- "cpu_ram_mb": 16757.338112,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
  "platform": "Linux-6.5.0-1023-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.3.1",
83
- "optimum_benchmark_commit": "c31e63ddb94c29460ba90ff599147e94ff80543a",
84
  "transformers_version": "4.42.4",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.32.1",
@@ -99,7 +100,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 1184.661504,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -108,37 +109,37 @@
108
  "latency": {
109
  "unit": "s",
110
  "count": 7,
111
- "total": 1.0379026440000985,
112
- "mean": 0.14827180628572836,
113
- "stdev": 0.010918568846618038,
114
- "p50": 0.15153902000002972,
115
- "p90": 0.15978533520001292,
116
- "p95": 0.16087156859998686,
117
- "p99": 0.161740555319966,
118
  "values": [
119
- 0.15778176300000268,
120
- 0.15833702400004768,
121
- 0.1619578019999608,
122
- 0.15153902000002972,
123
- 0.13613779500002465,
124
- 0.1360787830000163,
125
- 0.13607045700001663
126
  ]
127
  },
128
  "throughput": {
129
  "unit": "samples/s",
130
- "value": 6.744370524986673
131
  },
132
  "energy": {
133
  "unit": "kWh",
134
- "cpu": 5.3407708803812665e-06,
135
- "ram": 2.2320598338611813e-07,
136
  "gpu": 0.0,
137
- "total": 5.563976863767385e-06
138
  },
139
  "efficiency": {
140
  "unit": "samples/kWh",
141
- "value": 179727.5625842371
142
  }
143
  }
144
  }
 
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
10
+ "model_type": "deberta-v2",
11
  "model": "microsoft/deberta-v3-base",
12
  "processor": "microsoft/deberta-v3-base",
13
  "device": "cpu",
 
74
  "environment": {
75
  "cpu": " AMD EPYC 7763 64-Core Processor",
76
  "cpu_count": 4,
77
+ "cpu_ram_mb": 16757.342208,
78
  "system": "Linux",
79
  "machine": "x86_64",
80
  "platform": "Linux-6.5.0-1023-azure-x86_64-with-glibc2.35",
81
  "processor": "x86_64",
82
  "python_version": "3.10.14",
83
  "optimum_benchmark_version": "0.3.1",
84
+ "optimum_benchmark_commit": "402644384abc46743e46dac65be6512bc68c9d92",
85
  "transformers_version": "4.42.4",
86
  "transformers_commit": null,
87
  "accelerate_version": "0.32.1",
 
100
  "forward": {
101
  "memory": {
102
  "unit": "MB",
103
+ "max_ram": 1177.509888,
104
  "max_global_vram": null,
105
  "max_process_vram": null,
106
  "max_reserved": null,
 
109
  "latency": {
110
  "unit": "s",
111
  "count": 7,
112
+ "total": 1.0301348599999756,
113
+ "mean": 0.14716212285713937,
114
+ "stdev": 0.007939894131583141,
115
+ "p50": 0.15229217699999253,
116
+ "p90": 0.154878854399999,
117
+ "p95": 0.15604602570001588,
118
+ "p99": 0.1569797627400294,
119
  "values": [
120
+ 0.15721319700003278,
121
+ 0.15284159499998395,
122
+ 0.15229217699999253,
123
+ 0.15332262599997648,
124
+ 0.1386458169999969,
125
+ 0.13787474299999758,
126
+ 0.13794470499999534
127
  ]
128
  },
129
  "throughput": {
130
  "unit": "samples/s",
131
+ "value": 6.795226791956313
132
  },
133
  "energy": {
134
  "unit": "kWh",
135
+ "cpu": 5.404794050587549e-06,
136
+ "ram": 2.2588158137153867e-07,
137
  "gpu": 0.0,
138
+ "total": 5.630675631959088e-06
139
  },
140
  "efficiency": {
141
  "unit": "samples/kWh",
142
+ "value": 177598.58059024237
143
  }
144
  }
145
  }