elliesleightholm
commited on
Update app.py
Browse files
app.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import gradio as gr
|
2 |
import pandas as pd
|
3 |
|
4 |
-
# ---------------
|
5 |
|
6 |
-
# Embedding Models for
|
7 |
embedding_models = [
|
8 |
'<a href="https://docs.aws.amazon.com/bedrock/latest/userguide/titan-multiemb-models.html">Amazon-Titan-MultiModal</a>',
|
9 |
'<a href="https://huggingface.co/jinaai/jina-clip-v1">Jina-V1-CLIP</a>',
|
@@ -140,7 +140,7 @@ with gr.Blocks(css="""
|
|
140 |
gr.Markdown("# Ecommerce Embedding Model Benchmarks")
|
141 |
|
142 |
gr.Markdown("This Space contains benchmark results conducted as part of the release of our ecommerce embedding models: [**`Marqo-Ecommerce-L`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-L) and [**`Marqo-Ecommerce-B`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-B). ")
|
143 |
-
gr.Markdown(
|
144 |
gr.Markdown('Within both these scenarios, the models were benchmarked against three different tasks:')
|
145 |
gr.Markdown('- **Google Shopping Text-to-Image**')
|
146 |
gr.Markdown('- **Google Shopping Category-to-Image**')
|
@@ -160,7 +160,7 @@ with gr.Blocks(css="""
|
|
160 |
gr.Dataframe(value=hard_ap_text2_image_3m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
|
161 |
|
162 |
# Easy
|
163 |
-
gr.Markdown('##
|
164 |
gr.Markdown('### Google Shopping Text to Image')
|
165 |
gr.Dataframe(value=gs_text2_image_1m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
|
166 |
|
|
|
1 |
import gradio as gr
|
2 |
import pandas as pd
|
3 |
|
4 |
+
# --------------- Marqo-Ecommerce-Easy -------------------
|
5 |
|
6 |
+
# Embedding Models for Marqo-Ecommerce-Easy
|
7 |
embedding_models = [
|
8 |
'<a href="https://docs.aws.amazon.com/bedrock/latest/userguide/titan-multiemb-models.html">Amazon-Titan-MultiModal</a>',
|
9 |
'<a href="https://huggingface.co/jinaai/jina-clip-v1">Jina-V1-CLIP</a>',
|
|
|
140 |
gr.Markdown("# Ecommerce Embedding Model Benchmarks")
|
141 |
|
142 |
gr.Markdown("This Space contains benchmark results conducted as part of the release of our ecommerce embedding models: [**`Marqo-Ecommerce-L`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-L) and [**`Marqo-Ecommerce-B`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-B). ")
|
143 |
+
gr.Markdown("Our benchmarking process was divided into two distinct regimes, each using different datasets of ecommerce product listings: **marqo-ecommerce-hard** and **marqo-ecommerce-easy**. Both datasets contained product images and text and only differed in size. The "easy" dataset is approximately 10-30 times smaller (200k vs 4M products), and designed to accommodate rate-limited models, specifically Cohere-Embeddings-v3 and GCP-Vertex (with limits of 0.66 rps and 2 rps respectively). The "hard" dataset represents the true challenge, since it contains four million ecommerce product listings and is more representative of real-world ecommerce search scenarios.")
|
144 |
gr.Markdown('Within both these scenarios, the models were benchmarked against three different tasks:')
|
145 |
gr.Markdown('- **Google Shopping Text-to-Image**')
|
146 |
gr.Markdown('- **Google Shopping Category-to-Image**')
|
|
|
160 |
gr.Dataframe(value=hard_ap_text2_image_3m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
|
161 |
|
162 |
# Easy
|
163 |
+
gr.Markdown('## Marqo-Ecommerce-Easy')
|
164 |
gr.Markdown('### Google Shopping Text to Image')
|
165 |
gr.Dataframe(value=gs_text2_image_1m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
|
166 |
|