Spaces:
Running
Running
BenchmarkBot
commited on
Commit
β’
df1a500
1
Parent(s):
fbbd324
fix about
Browse files- app.py +1 -1
- src/assets/text_content.py +7 -5
app.py
CHANGED
@@ -310,7 +310,7 @@ with demo:
|
|
310 |
)
|
311 |
|
312 |
with gr.TabItem("β About π", id=4):
|
313 |
-
gr.
|
314 |
|
315 |
demo.load(
|
316 |
change_tab,
|
|
|
310 |
)
|
311 |
|
312 |
with gr.TabItem("β About π", id=4):
|
313 |
+
gr.HTML(ABOUT_TEXT)
|
314 |
|
315 |
demo.load(
|
316 |
change_tab,
|
src/assets/text_content.py
CHANGED
@@ -16,11 +16,13 @@ A100_TEXT = """<h3>Single-GPU Benchmark (1xA100):</h3>
|
|
16 |
"""
|
17 |
|
18 |
ABOUT_TEXT = """<h3>About the benchmarks:</h3>
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
|
|
|
|
24 |
"""
|
25 |
|
26 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results."
|
|
|
16 |
"""
|
17 |
|
18 |
ABOUT_TEXT = """<h3>About the benchmarks:</h3>
|
19 |
+
<ul>
|
20 |
+
<li>The performances benchmarks were obtained using <a href="https://github.com/huggingface/optimum-benchmark">Optimum-Benchmark</a>.</li>
|
21 |
+
<li>Throughput is measured in tokens per second when generating 1000 tokens with a batch size of 1.</li>
|
22 |
+
<li>Peak memory is measured in MB during the first forward pass of the model (no warmup).</li>
|
23 |
+
<li>Open LLM Score is an average evaluation score obtained from the <a href="https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard">π€ Open LLM Leaderboard</a>.</li>
|
24 |
+
<li>Open LLM Tradeoff is the euclidean distance between an LLM and the "perfect LLM" (i.e. 0 latency and 100% accuracy) translating the tradeoff between latency and accuracy.</li>
|
25 |
+
</ul>
|
26 |
"""
|
27 |
|
28 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results."
|