elliesleightholm commited on
Commit
8b7c6a9
·
verified ·
1 Parent(s): ae4d4ac

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -1,9 +1,9 @@
1
  import gradio as gr
2
  import pandas as pd
3
 
4
- # --------------- 100k-Marqo-Ecommerce-Easy -------------------
5
 
6
- # Embedding Models for 100k-Marqo-Ecommerce-Easy
7
  embedding_models = [
8
  '<a href="https://docs.aws.amazon.com/bedrock/latest/userguide/titan-multiemb-models.html">Amazon-Titan-MultiModal</a>',
9
  '<a href="https://huggingface.co/jinaai/jina-clip-v1">Jina-V1-CLIP</a>',
@@ -140,7 +140,7 @@ with gr.Blocks(css="""
140
  gr.Markdown("# Ecommerce Embedding Model Benchmarks")
141
 
142
  gr.Markdown("This Space contains benchmark results conducted as part of the release of our ecommerce embedding models: [**`Marqo-Ecommerce-L`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-L) and [**`Marqo-Ecommerce-B`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-B). ")
143
- gr.Markdown('The benchmarks are separated into \'Marqo-Ecommerce-Hard\' and \'100k-Marqo-Ecommerce-Easy\'. The "easy" dataset is about 10-30 times smaller, and designed to accommodate rate-limited models, specifically Cohere-Embeddings-v3 and GCP-Vertex. The "hard" dataset represents the true challenge, since it contains four million ecommerce product listings, which pushes these models to their limits in a real-world, ecommerce scenario. ' )
144
  gr.Markdown('Within both these scenarios, the models were benchmarked against three different tasks:')
145
  gr.Markdown('- **Google Shopping Text-to-Image**')
146
  gr.Markdown('- **Google Shopping Category-to-Image**')
@@ -160,7 +160,7 @@ with gr.Blocks(css="""
160
  gr.Dataframe(value=hard_ap_text2_image_3m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
161
 
162
  # Easy
163
- gr.Markdown('## 100k-Marqo-Ecommerce-Easy')
164
  gr.Markdown('### Google Shopping Text to Image')
165
  gr.Dataframe(value=gs_text2_image_1m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
166
 
 
1
  import gradio as gr
2
  import pandas as pd
3
 
4
+ # --------------- Marqo-Ecommerce-Easy -------------------
5
 
6
+ # Embedding Models for Marqo-Ecommerce-Easy
7
  embedding_models = [
8
  '<a href="https://docs.aws.amazon.com/bedrock/latest/userguide/titan-multiemb-models.html">Amazon-Titan-MultiModal</a>',
9
  '<a href="https://huggingface.co/jinaai/jina-clip-v1">Jina-V1-CLIP</a>',
 
140
  gr.Markdown("# Ecommerce Embedding Model Benchmarks")
141
 
142
  gr.Markdown("This Space contains benchmark results conducted as part of the release of our ecommerce embedding models: [**`Marqo-Ecommerce-L`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-L) and [**`Marqo-Ecommerce-B`**](https://huggingface.co/Marqo/marqo-ecommerce-embeddings-B). ")
143
+ gr.Markdown("Our benchmarking process was divided into two distinct regimes, each using different datasets of ecommerce product listings: **marqo-ecommerce-hard** and **marqo-ecommerce-easy**. Both datasets contained product images and text and only differed in size. The "easy" dataset is approximately 10-30 times smaller (200k vs 4M products), and designed to accommodate rate-limited models, specifically Cohere-Embeddings-v3 and GCP-Vertex (with limits of 0.66 rps and 2 rps respectively). The "hard" dataset represents the true challenge, since it contains four million ecommerce product listings and is more representative of real-world ecommerce search scenarios.")
144
  gr.Markdown('Within both these scenarios, the models were benchmarked against three different tasks:')
145
  gr.Markdown('- **Google Shopping Text-to-Image**')
146
  gr.Markdown('- **Google Shopping Category-to-Image**')
 
160
  gr.Dataframe(value=hard_ap_text2_image_3m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
161
 
162
  # Easy
163
+ gr.Markdown('## Marqo-Ecommerce-Easy')
164
  gr.Markdown('### Google Shopping Text to Image')
165
  gr.Dataframe(value=gs_text2_image_1m_df, headers="keys", interactive=True, datatype=["html", "number", "number", "number", "number"])
166