Upload folder using huggingface_hub
#2
by
begumcig
- opened
- base_results.json +19 -0
- plots.png +0 -0
- smashed_results.json +19 -0
base_results.json
ADDED
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"current_gpu_type": "Tesla T4",
|
3 |
+
"current_gpu_total_memory": 15095.0625,
|
4 |
+
"perplexity": 2.7025578022003174,
|
5 |
+
"memory_inference_first": 1710.0,
|
6 |
+
"memory_inference": 1710.0,
|
7 |
+
"token_generation_latency_sync": 43.00895957946777,
|
8 |
+
"token_generation_latency_async": 42.318619787693024,
|
9 |
+
"token_generation_throughput_sync": 0.02325096932773501,
|
10 |
+
"token_generation_throughput_async": 0.023630260273536072,
|
11 |
+
"token_generation_CO2_emissions": 4.656282724561804e-05,
|
12 |
+
"token_generation_energy_consumption": 0.002189405360918023,
|
13 |
+
"inference_latency_sync": 278.88348388671875,
|
14 |
+
"inference_latency_async": 101.10292434692383,
|
15 |
+
"inference_throughput_sync": 0.003585726863646739,
|
16 |
+
"inference_throughput_async": 0.009890910737345316,
|
17 |
+
"inference_CO2_emissions": 4.665015602960695e-05,
|
18 |
+
"inference_energy_consumption": 0.00016071153594523843
|
19 |
+
}
|
plots.png
ADDED
smashed_results.json
ADDED
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"current_gpu_type": "Tesla T4",
|
3 |
+
"current_gpu_total_memory": 15095.0625,
|
4 |
+
"perplexity": 2.8998236656188965,
|
5 |
+
"memory_inference_first": 358.0,
|
6 |
+
"memory_inference": 358.0,
|
7 |
+
"token_generation_latency_sync": 80.11734771728516,
|
8 |
+
"token_generation_latency_async": 80.36270011216402,
|
9 |
+
"token_generation_throughput_sync": 0.012481691275262373,
|
10 |
+
"token_generation_throughput_async": 0.012443583884118847,
|
11 |
+
"token_generation_CO2_emissions": 2.351045754117281e-05,
|
12 |
+
"token_generation_energy_consumption": 0.005466914504574626,
|
13 |
+
"inference_latency_sync": 79.62785682678222,
|
14 |
+
"inference_latency_async": 59.72108840942383,
|
15 |
+
"inference_throughput_sync": 0.012558419124293918,
|
16 |
+
"inference_throughput_async": 0.016744503937108465,
|
17 |
+
"inference_CO2_emissions": 2.3351902662235743e-05,
|
18 |
+
"inference_energy_consumption": 8.117517466477338e-05
|
19 |
+
}
|