prithivida commited on
Commit
47cf1a0
1 Parent(s): 5f315c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -55,13 +55,13 @@ Learned Neural / Dense retrievers (DPR, Sentence transformers*, BGE* models) wit
55
 
56
  **3. The big idea:**
57
 
58
- Getting pros of both searches made sense and that gave rise to interest in learning sparse representations for queries and documents with some interpretability. The sparse representations also double as implicit or explicit (latent, contextualized) expansion mechanisms for both query and documents. If you are new to query expansion learn more here from the master himself Daniel Tunkelang (link below).
59
 
60
  **4. What a Sparse model learns ?**
61
 
62
  The model learns to project it's learned dense representations over a MLM head to give a vocabulary distribution. Which is just to say the model can do automatic token expansion. (Image courtesy of pinecone)
63
 
64
- <img src="./expansion.png" width=650 height=500/>
65
 
66
  </details>
67
 
@@ -76,7 +76,7 @@ SPLADE models are a fine balance between retrieval effectiveness (quality) and r
76
  4. Achieves a modest yet competitive effectiveness **MRR@10 37.22** in ID data (& OOD) and a retrieval latency of - **47.27ms**. (multi-threaded) all On **Consumer grade-GPUs** with **only 5 negatives per query**.
77
  4. For Industry setting: Effectiveness on custom domains needs more than just **Trading FLOPS for tiny gains** and The Premise "SPLADE++ are not well suited to mono-cpu retrieval" does not hold.
78
 
79
- <img src="./ID.png" width=650 height=500/>
80
 
81
  *Note: The paper refers to the best performing models as SPLADE++, hence for consistency we are reusing the same.*
82
 
 
55
 
56
  **3. The big idea:**
57
 
58
+ Getting pros of both searches made sense and that gave rise to interest in learning sparse representations for queries and documents with some interpretability. The sparse representations also double as implicit or explicit (latent, contextualized) expansion mechanisms for both query and documents. If you are new to query expansion learn more here from the master himself Daniel Tunkelang.
59
 
60
  **4. What a Sparse model learns ?**
61
 
62
  The model learns to project it's learned dense representations over a MLM head to give a vocabulary distribution. Which is just to say the model can do automatic token expansion. (Image courtesy of pinecone)
63
 
64
+ <img src="./expansion.png" width=600 height=550/>
65
 
66
  </details>
67
 
 
76
  4. Achieves a modest yet competitive effectiveness **MRR@10 37.22** in ID data (& OOD) and a retrieval latency of - **47.27ms**. (multi-threaded) all On **Consumer grade-GPUs** with **only 5 negatives per query**.
77
  4. For Industry setting: Effectiveness on custom domains needs more than just **Trading FLOPS for tiny gains** and The Premise "SPLADE++ are not well suited to mono-cpu retrieval" does not hold.
78
 
79
+ <img src="./ID.png" width=750 height=650/>
80
 
81
  *Note: The paper refers to the best performing models as SPLADE++, hence for consistency we are reusing the same.*
82