Upload folder using huggingface_hub

#2
by begumcig - opened
Files changed (3) hide show
  1. base_results.json +19 -0
  2. plots.png +0 -0
  3. smashed_results.json +19 -0
base_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 2.7025578022003174,
5
+ "memory_inference_first": 1710.0,
6
+ "memory_inference": 1710.0,
7
+ "token_generation_latency_sync": 43.00895957946777,
8
+ "token_generation_latency_async": 42.318619787693024,
9
+ "token_generation_throughput_sync": 0.02325096932773501,
10
+ "token_generation_throughput_async": 0.023630260273536072,
11
+ "token_generation_CO2_emissions": 4.656282724561804e-05,
12
+ "token_generation_energy_consumption": 0.002189405360918023,
13
+ "inference_latency_sync": 278.88348388671875,
14
+ "inference_latency_async": 101.10292434692383,
15
+ "inference_throughput_sync": 0.003585726863646739,
16
+ "inference_throughput_async": 0.009890910737345316,
17
+ "inference_CO2_emissions": 4.665015602960695e-05,
18
+ "inference_energy_consumption": 0.00016071153594523843
19
+ }
plots.png ADDED
smashed_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 2.8998236656188965,
5
+ "memory_inference_first": 358.0,
6
+ "memory_inference": 358.0,
7
+ "token_generation_latency_sync": 80.11734771728516,
8
+ "token_generation_latency_async": 80.36270011216402,
9
+ "token_generation_throughput_sync": 0.012481691275262373,
10
+ "token_generation_throughput_async": 0.012443583884118847,
11
+ "token_generation_CO2_emissions": 2.351045754117281e-05,
12
+ "token_generation_energy_consumption": 0.005466914504574626,
13
+ "inference_latency_sync": 79.62785682678222,
14
+ "inference_latency_async": 59.72108840942383,
15
+ "inference_throughput_sync": 0.012558419124293918,
16
+ "inference_throughput_async": 0.016744503937108465,
17
+ "inference_CO2_emissions": 2.3351902662235743e-05,
18
+ "inference_energy_consumption": 8.117517466477338e-05
19
+ }