Upload folder using huggingface_hub

#2
by begumcig - opened
Files changed (3) hide show
  1. base_results.json +19 -0
  2. plots.png +0 -0
  3. smashed_results.json +19 -0
base_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 2.2062017917633057,
5
+ "memory_inference_first": 7278.0,
6
+ "memory_inference": 7278.0,
7
+ "token_generation_latency_sync": 35.36079788208008,
8
+ "token_generation_latency_async": 34.97263416647911,
9
+ "token_generation_throughput_sync": 0.0282799048634243,
10
+ "token_generation_throughput_async": 0.028593785507826836,
11
+ "token_generation_CO2_emissions": 0.00018337661498006707,
12
+ "token_generation_energy_consumption": 0.002160617961299883,
13
+ "inference_latency_sync": 1060.47254486084,
14
+ "inference_latency_async": 136.61155700683594,
15
+ "inference_throughput_sync": 0.000942975850573505,
16
+ "inference_throughput_async": 0.007320024907921669,
17
+ "inference_CO2_emissions": 0.0001840978910408864,
18
+ "inference_energy_consumption": 0.0006337665377060597
19
+ }
plots.png ADDED
smashed_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 30815030.0,
5
+ "memory_inference_first": 1142.0,
6
+ "memory_inference": 1142.0,
7
+ "token_generation_latency_sync": 147.6204864501953,
8
+ "token_generation_latency_async": 147.56362289190292,
9
+ "token_generation_throughput_sync": 0.006774127521503482,
10
+ "token_generation_throughput_async": 0.006776737927697435,
11
+ "token_generation_CO2_emissions": 0.00020048614177282614,
12
+ "token_generation_energy_consumption": 0.008781923951582248,
13
+ "inference_latency_sync": 1151.6722137451172,
14
+ "inference_latency_async": 424.6630907058716,
15
+ "inference_throughput_sync": 0.0008683026194997837,
16
+ "inference_throughput_async": 0.00235480789803938,
17
+ "inference_CO2_emissions": 0.0002004475962713345,
18
+ "inference_energy_consumption": 0.0006890801651842688
19
+ }