IlyasMoutawwakil HF staff commited on
Commit
57aaac0
·
verified ·
1 Parent(s): daabddd

Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark_report.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "forward": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 904.64256,
6
+ "max_global_vram": 1057.488896,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 406.847488,
9
+ "max_allocated": 364.432384
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 169,
14
+ "total": 0.995434849739075,
15
+ "mean": 0.0058901470398761814,
16
+ "stdev": 0.0002672380576739546,
17
+ "p50": 0.005928959846496582,
18
+ "p90": 0.006132121753692627,
19
+ "p95": 0.006424575901031494,
20
+ "p99": 0.006725652408599853,
21
+ "values": [
22
+ 0.006331391811370849,
23
+ 0.006081535816192627,
24
+ 0.00609388780593872,
25
+ 0.006128640174865723,
26
+ 0.006146048069000244,
27
+ 0.006072319984436035,
28
+ 0.006139904022216797,
29
+ 0.0061931519508361815,
30
+ 0.0060590081214904785,
31
+ 0.0060897917747497554,
32
+ 0.006051839828491211,
33
+ 0.0061931519508361815,
34
+ 0.006067200183868408,
35
+ 0.005976064205169678,
36
+ 0.00607539176940918,
37
+ 0.005917695999145508,
38
+ 0.005955584049224853,
39
+ 0.005915647983551026,
40
+ 0.005985280036926269,
41
+ 0.005988319873809814,
42
+ 0.006052864074707031,
43
+ 0.006045695781707764,
44
+ 0.006106175899505615,
45
+ 0.006065152168273926,
46
+ 0.006049791812896729,
47
+ 0.0059391999244689945,
48
+ 0.006004735946655273,
49
+ 0.0059699201583862304,
50
+ 0.006060031890869141,
51
+ 0.006017024040222168,
52
+ 0.005982272148132324,
53
+ 0.006030335903167725,
54
+ 0.006039552211761475,
55
+ 0.005997568130493164,
56
+ 0.00601801586151123,
57
+ 0.005989376068115234,
58
+ 0.006032383918762207,
59
+ 0.0062197761535644535,
60
+ 0.0059556479454040525,
61
+ 0.005941247940063477,
62
+ 0.0065372161865234375,
63
+ 0.006696959972381592,
64
+ 0.0069407038688659665,
65
+ 0.005883903980255127,
66
+ 0.005975039958953857,
67
+ 0.005992447853088379,
68
+ 0.005991424083709717,
69
+ 0.005997568130493164,
70
+ 0.006054912090301514,
71
+ 0.006001664161682129,
72
+ 0.006051839828491211,
73
+ 0.0060293121337890625,
74
+ 0.006031360149383545,
75
+ 0.005946368217468262,
76
+ 0.0067010560035705566,
77
+ 0.005983232021331787,
78
+ 0.006137856006622314,
79
+ 0.006415359973907471,
80
+ 0.00643071985244751,
81
+ 0.006037504196166992,
82
+ 0.005910528182983398,
83
+ 0.005956607818603516,
84
+ 0.005955584049224853,
85
+ 0.0059955201148986816,
86
+ 0.005952511787414551,
87
+ 0.006097919940948486,
88
+ 0.0059391999244689945,
89
+ 0.005928959846496582,
90
+ 0.006130688190460205,
91
+ 0.006007808208465576,
92
+ 0.005925888061523437,
93
+ 0.005963776111602783,
94
+ 0.005993472099304199,
95
+ 0.005954592227935791,
96
+ 0.005914624214172363,
97
+ 0.006021120071411133,
98
+ 0.00601801586151123,
99
+ 0.006054912090301514,
100
+ 0.006044672012329101,
101
+ 0.0060405759811401364,
102
+ 0.005979135990142822,
103
+ 0.006008831977844238,
104
+ 0.005946368217468262,
105
+ 0.006011903762817383,
106
+ 0.005847040176391601,
107
+ 0.005990399837493897,
108
+ 0.005911551952362061,
109
+ 0.00591974401473999,
110
+ 0.005818367958068848,
111
+ 0.0058039999008178715,
112
+ 0.005800000190734863,
113
+ 0.005973055839538575,
114
+ 0.005788671970367432,
115
+ 0.005626880168914795,
116
+ 0.0056258559226989744,
117
+ 0.0057077760696411135,
118
+ 0.005756927967071533,
119
+ 0.0057825279235839844,
120
+ 0.005823488235473633,
121
+ 0.005775360107421875,
122
+ 0.005900288105010986,
123
+ 0.005953536033630371,
124
+ 0.006777919769287109,
125
+ 0.006627327919006347,
126
+ 0.0065413122177124024,
127
+ 0.006533120155334473,
128
+ 0.0057149438858032225,
129
+ 0.005634047985076904,
130
+ 0.00568012809753418,
131
+ 0.005710847854614258,
132
+ 0.0056145920753479005,
133
+ 0.005641215801239013,
134
+ 0.005662720203399658,
135
+ 0.005610496044158936,
136
+ 0.005640192031860352,
137
+ 0.00566374397277832,
138
+ 0.005628928184509278,
139
+ 0.005591040134429932,
140
+ 0.005596159934997558,
141
+ 0.005627903938293457,
142
+ 0.00561356782913208,
143
+ 0.005626880168914795,
144
+ 0.005599232196807862,
145
+ 0.005659647941589355,
146
+ 0.005642240047454834,
147
+ 0.005620736122131348,
148
+ 0.005616640090942383,
149
+ 0.0056258559226989744,
150
+ 0.005618688106536865,
151
+ 0.005627903938293457,
152
+ 0.005604320049285889,
153
+ 0.005623807907104492,
154
+ 0.005615615844726562,
155
+ 0.005635072231292724,
156
+ 0.005609471797943116,
157
+ 0.005584832191467285,
158
+ 0.005641215801239013,
159
+ 0.005612576007843017,
160
+ 0.005631008148193359,
161
+ 0.005617663860321045,
162
+ 0.005633024215698243,
163
+ 0.005699584007263184,
164
+ 0.005607359886169434,
165
+ 0.005626880168914795,
166
+ 0.005606400012969971,
167
+ 0.00561356782913208,
168
+ 0.005643263816833496,
169
+ 0.005637119770050049,
170
+ 0.005616640090942383,
171
+ 0.005608448028564453,
172
+ 0.0056442880630493165,
173
+ 0.00562278413772583,
174
+ 0.005599232196807862,
175
+ 0.005620736122131348,
176
+ 0.005610496044158936,
177
+ 0.0056217598915100095,
178
+ 0.00561356782913208,
179
+ 0.005610496044158936,
180
+ 0.005659647941589355,
181
+ 0.005667840003967285,
182
+ 0.005629951953887939,
183
+ 0.005616640090942383,
184
+ 0.005637119770050049,
185
+ 0.005636096000671387,
186
+ 0.0056442880630493165,
187
+ 0.0056217598915100095,
188
+ 0.005595136165618897,
189
+ 0.005599232196807862,
190
+ 0.005594111919403076
191
+ ]
192
+ },
193
+ "throughput": {
194
+ "unit": "samples/s",
195
+ "value": 169.7750486074489
196
+ },
197
+ "energy": {
198
+ "unit": "kWh",
199
+ "cpu": 6.491535981640935e-08,
200
+ "ram": 3.5490031235089874e-08,
201
+ "gpu": 3.6680964378142444e-07,
202
+ "total": 4.6721503483292366e-07
203
+ },
204
+ "efficiency": {
205
+ "unit": "samples/kWh",
206
+ "value": 2140342.0811524196
207
+ }
208
+ }
209
+ }