IlyasMoutawwakil HF staff commited on
Commit
732318b
·
verified ·
1 Parent(s): c0e73e8

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json CHANGED
@@ -2,187 +2,190 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 878.587904,
6
- "max_global_vram": 1231.552512,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 591.396864,
9
  "max_allocated": 449.269248
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 148,
14
- "total": 1.0059274244308472,
15
- "mean": 0.006796806921830048,
16
- "stdev": 0.00046711873956685685,
17
- "p50": 0.0067066879272460935,
18
- "p90": 0.00689469428062439,
19
- "p95": 0.0071596033573150634,
20
- "p99": 0.009348465013504028,
21
  "values": [
22
- 0.009962495803833007,
23
- 0.009354240417480468,
24
- 0.009341952323913574,
25
- 0.009155584335327148,
26
- 0.007359488010406494,
27
- 0.007200767993927002,
28
  0.0072325119972229,
29
- 0.007163904190063477,
30
- 0.007072768211364746,
31
- 0.00667852783203125,
32
- 0.006730751991271973,
33
- 0.006702079772949219,
34
- 0.006622208118438721,
35
- 0.006604800224304199,
36
- 0.006627327919006347,
37
- 0.006654975891113281,
38
- 0.006713344097137451,
39
- 0.006755328178405762,
40
- 0.006762495994567871,
41
- 0.006763519763946534,
42
- 0.006802432060241699,
43
- 0.006721536159515381,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  0.006688767910003662,
45
- 0.006716415882110595,
46
- 0.006792191982269287,
47
- 0.006731776237487793,
48
- 0.006770688056945801,
49
- 0.006782976150512696,
50
- 0.006761472225189209,
51
- 0.0067041277885437015,
52
- 0.006716415882110595,
53
- 0.0067041277885437015,
54
- 0.006758399963378906,
55
- 0.006718463897705078,
56
- 0.006625279903411865,
57
  0.006619135856628418,
58
- 0.00662937593460083,
59
- 0.0066447358131408694,
60
  0.006660096168518067,
61
- 0.006899712085723877,
62
- 0.006708223819732666,
63
- 0.006719488143920899,
64
- 0.006748159885406494,
65
- 0.006726655960083008,
66
- 0.006774784088134766,
67
- 0.006781951904296875,
68
- 0.006774784088134766,
69
- 0.006688767910003662,
70
- 0.006670335769653321,
71
- 0.006656000137329102,
72
- 0.0066979842185974124,
73
- 0.0067358717918396,
74
- 0.0066979842185974124,
75
- 0.006747136116027832,
76
- 0.006723584175109864,
77
- 0.006606847763061524,
78
- 0.006633471965789795,
79
- 0.00667852783203125,
80
- 0.006710271835327148,
81
- 0.006779903888702392,
82
- 0.006817791938781738,
83
- 0.0067010560035705566,
84
- 0.00659660816192627,
85
- 0.006692863941192627,
86
- 0.006687744140625,
87
- 0.0066826238632202144,
88
- 0.006658048152923584,
89
- 0.0066713600158691405,
90
- 0.00667852783203125,
91
- 0.006756351947784424,
92
- 0.006782976150512696,
93
- 0.006719488143920899,
94
- 0.00678604793548584,
95
- 0.006782976150512696,
96
- 0.006763519763946534,
97
- 0.006672383785247803,
98
- 0.006772736072540283,
99
  0.006681600093841553,
100
- 0.006650879859924317,
101
- 0.006797311782836914,
102
- 0.006604800224304199,
103
- 0.006516736030578613,
104
- 0.006471680164337158,
105
- 0.0064174079895019534,
106
- 0.006529024124145508,
107
- 0.006504447937011719,
108
- 0.006384640216827392,
109
- 0.006371327877044678,
110
- 0.006362112045288086,
111
- 0.0064102401733398436,
112
- 0.006862847805023193,
113
- 0.006870016098022461,
114
- 0.006890495777130127,
115
- 0.006909952163696289,
116
- 0.006999040126800537,
117
- 0.007018496036529541,
118
- 0.007151616096496582,
119
- 0.007030784130096435,
120
- 0.006795263767242431,
121
- 0.006729728221893311,
122
- 0.006792191982269287,
123
- 0.00678604793548584,
124
  0.0067010560035705566,
125
- 0.006724607944488525,
126
- 0.006759424209594727,
127
- 0.0067051520347595215,
128
- 0.006708223819732666,
129
- 0.006708223819732666,
130
- 0.0067051520347595215,
131
  0.006656000137329102,
132
- 0.0067051520347595215,
133
- 0.006668288230895996,
134
- 0.006685696125030518,
135
- 0.006681600093841553,
136
- 0.006715392112731934,
137
- 0.00669593620300293,
138
- 0.006744063854217529,
139
- 0.0066344962120056155,
140
- 0.0066344962120056155,
141
- 0.006624256134033203,
142
- 0.0067041277885437015,
143
- 0.0066938881874084475,
144
- 0.006686719894409179,
145
- 0.006690815925598144,
146
- 0.006673408031463623,
147
- 0.006718463897705078,
148
- 0.006718463897705078,
149
- 0.00667852783203125,
150
- 0.006759424209594727,
151
  0.00669593620300293,
152
- 0.006688767910003662,
153
- 0.006714367866516113,
154
- 0.006677504062652588,
155
- 0.006643712043762207,
156
- 0.006892543792724609,
157
- 0.0067123198509216305,
158
  0.006654975891113281,
159
- 0.006672383785247803,
160
- 0.006726655960083008,
161
- 0.006635519981384277,
162
- 0.0067041277885437015,
163
- 0.0067123198509216305,
164
- 0.006650879859924317,
165
  0.00667852783203125,
166
- 0.006710271835327148,
167
- 0.006657023906707763,
168
- 0.006690815925598144,
169
- 0.0066979842185974124
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
170
  ]
171
  },
172
  "throughput": {
173
  "unit": "samples/s",
174
- "value": 147.1279104292621
175
  },
176
  "energy": {
177
  "unit": "kWh",
178
- "cpu": 7.811371585604741e-08,
179
- "ram": 4.270364973691405e-08,
180
- "gpu": 1.5042592443421047e-07,
181
- "total": 2.712432900271719e-07
182
  },
183
  "efficiency": {
184
  "unit": "samples/kWh",
185
- "value": 3686727.1441067704
186
  }
187
  }
188
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 877.416448,
6
+ "max_global_vram": 1229.45536,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 591.396864,
9
  "max_allocated": 449.269248
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 151,
14
+ "total": 1.0013317770957948,
15
+ "mean": 0.00663133627215758,
16
+ "stdev": 0.00032869613162828185,
17
+ "p50": 0.006593535900115967,
18
+ "p90": 0.007223296165466309,
19
+ "p95": 0.007343616008758545,
20
+ "p99": 0.007781376123428345,
21
  "values": [
22
+ 0.007622655868530274,
23
+ 0.007482367992401123,
24
+ 0.0073768959045410155,
 
 
 
25
  0.0072325119972229,
26
+ 0.007310336112976074,
27
+ 0.00728985595703125,
28
+ 0.007223296165466309,
29
+ 0.007271423816680909,
30
+ 0.0072468481063842774,
31
+ 0.007265279769897461,
32
+ 0.007175168037414551,
33
+ 0.007292928218841553,
34
+ 0.007498752117156982,
35
+ 0.006545407772064209,
36
+ 0.006367231845855713,
37
+ 0.006322175979614258,
38
+ 0.006306816101074219,
39
+ 0.006345727920532227,
40
+ 0.006345727920532227,
41
+ 0.00636518383026123,
42
+ 0.006304768085479737,
43
+ 0.006348800182342529,
44
+ 0.006325247764587402,
45
+ 0.006318079948425293,
46
+ 0.006341631889343262,
47
+ 0.006346752166748047,
48
+ 0.0063508481979370115,
49
+ 0.0063272957801818845,
50
+ 0.006311935901641846,
51
+ 0.006519807815551758,
52
+ 0.006335487842559814,
53
+ 0.0063508481979370115,
54
+ 0.006300672054290772,
55
+ 0.006325247764587402,
56
+ 0.006336512088775635,
57
+ 0.0063201279640197755,
58
+ 0.006340608119964599,
59
+ 0.006291456222534179,
60
+ 0.0063170561790466305,
61
+ 0.006296576023101807,
62
+ 0.006319104194641113,
63
+ 0.006296576023101807,
64
+ 0.006358016014099121,
65
+ 0.0063610877990722655,
66
+ 0.006302720069885254,
67
+ 0.006343679904937744,
68
+ 0.006338560104370118,
69
+ 0.006299647808074951,
70
+ 0.006316031932830811,
71
+ 0.006277120113372803,
72
+ 0.006339583873748779,
73
+ 0.006313983917236328,
74
+ 0.0066119680404663084,
75
+ 0.006821887969970703,
76
  0.006688767910003662,
77
+ 0.006658048152923584,
78
+ 0.00673689603805542,
79
+ 0.006702079772949219,
80
+ 0.006669312000274658,
81
+ 0.006710271835327148,
82
+ 0.006699007987976074,
83
+ 0.006647808074951172,
84
+ 0.006730751991271973,
85
+ 0.006667263984680176,
86
+ 0.006662144184112549,
87
+ 0.006681600093841553,
88
+ 0.0066744318008422855,
89
  0.006619135856628418,
90
+ 0.006663167953491211,
91
+ 0.006635519981384277,
92
  0.006660096168518067,
93
+ 0.0066641921997070315,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94
  0.006681600093841553,
95
+ 0.006646783828735352,
96
+ 0.006688767910003662,
97
+ 0.006668288230895996,
98
+ 0.006653952121734619,
99
+ 0.006700032234191895,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
100
  0.0067010560035705566,
101
+ 0.006662144184112549,
102
+ 0.006662144184112549,
103
+ 0.006653952121734619,
 
 
 
104
  0.006656000137329102,
105
+ 0.0066744318008422855,
106
+ 0.006657023906707763,
107
+ 0.006680575847625733,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  0.00669593620300293,
109
+ 0.006657023906707763,
110
+ 0.006638591766357422,
 
 
 
 
111
  0.006654975891113281,
112
+ 0.006667263984680176,
113
+ 0.006636544227600098,
114
+ 0.0066744318008422855,
115
+ 0.006642687797546387,
116
+ 0.006662144184112549,
 
117
  0.00667852783203125,
118
+ 0.006666240215301514,
119
+ 0.0066304001808166506,
120
+ 0.0066641921997070315,
121
+ 0.0066344962120056155,
122
+ 0.006649856090545654,
123
+ 0.006656000137329102,
124
+ 0.007621632099151611,
125
+ 0.007776256084442139,
126
+ 0.007786496162414551,
127
+ 0.007790592193603516,
128
+ 0.0064839677810668945,
129
+ 0.006482944011688232,
130
+ 0.006427648067474365,
131
+ 0.006391808032989502,
132
+ 0.006388735771179199,
133
+ 0.0063170561790466305,
134
+ 0.00642252779006958,
135
+ 0.0063170561790466305,
136
+ 0.006384640216827392,
137
+ 0.006339583873748779,
138
+ 0.006302720069885254,
139
+ 0.006303743839263916,
140
+ 0.006573056221008301,
141
+ 0.006557695865631104,
142
+ 0.006569983959197998,
143
+ 0.0065710082054138185,
144
+ 0.006589439868927002,
145
+ 0.0065474557876586915,
146
+ 0.00658739185333252,
147
+ 0.006593535900115967,
148
+ 0.006602752208709717,
149
+ 0.006549503803253174,
150
+ 0.006553599834442139,
151
+ 0.006582272052764892,
152
+ 0.006564864158630371,
153
+ 0.006557695865631104,
154
+ 0.006620160102844238,
155
+ 0.007139328002929687,
156
+ 0.006807551860809326,
157
+ 0.006626304149627686,
158
+ 0.006558720111846924,
159
+ 0.006564864158630371,
160
+ 0.0066109437942504885,
161
+ 0.006562816143035889,
162
+ 0.00657919979095459,
163
+ 0.006552576065063476,
164
+ 0.006620160102844238,
165
+ 0.006550528049468994,
166
+ 0.006559743881225586,
167
+ 0.006563839912414551,
168
+ 0.006560768127441406,
169
+ 0.006554624080657959,
170
+ 0.00667142391204834,
171
+ 0.006586368083953857,
172
+ 0.0065413122177124024
173
  ]
174
  },
175
  "throughput": {
176
  "unit": "samples/s",
177
+ "value": 150.79916912050047
178
  },
179
  "energy": {
180
  "unit": "kWh",
181
+ "cpu": 7.41030958791574e-08,
182
+ "ram": 4.050162078442554e-08,
183
+ "gpu": 1.4088205715000146e-07,
184
+ "total": 2.554867738135844e-07
185
  },
186
  "efficiency": {
187
  "unit": "samples/kWh",
188
+ "value": 3914096.941588251
189
  }
190
  }
191
  }