prithivida
commited on
Commit
•
47cf1a0
1
Parent(s):
5f315c6
Update README.md
Browse files
README.md
CHANGED
@@ -55,13 +55,13 @@ Learned Neural / Dense retrievers (DPR, Sentence transformers*, BGE* models) wit
|
|
55 |
|
56 |
**3. The big idea:**
|
57 |
|
58 |
-
Getting pros of both searches made sense and that gave rise to interest in learning sparse representations for queries and documents with some interpretability. The sparse representations also double as implicit or explicit (latent, contextualized) expansion mechanisms for both query and documents. If you are new to query expansion learn more here from the master himself Daniel Tunkelang
|
59 |
|
60 |
**4. What a Sparse model learns ?**
|
61 |
|
62 |
The model learns to project it's learned dense representations over a MLM head to give a vocabulary distribution. Which is just to say the model can do automatic token expansion. (Image courtesy of pinecone)
|
63 |
|
64 |
-
<img src="./expansion.png" width=
|
65 |
|
66 |
</details>
|
67 |
|
@@ -76,7 +76,7 @@ SPLADE models are a fine balance between retrieval effectiveness (quality) and r
|
|
76 |
4. Achieves a modest yet competitive effectiveness **MRR@10 37.22** in ID data (& OOD) and a retrieval latency of - **47.27ms**. (multi-threaded) all On **Consumer grade-GPUs** with **only 5 negatives per query**.
|
77 |
4. For Industry setting: Effectiveness on custom domains needs more than just **Trading FLOPS for tiny gains** and The Premise "SPLADE++ are not well suited to mono-cpu retrieval" does not hold.
|
78 |
|
79 |
-
<img src="./ID.png" width=
|
80 |
|
81 |
*Note: The paper refers to the best performing models as SPLADE++, hence for consistency we are reusing the same.*
|
82 |
|
|
|
55 |
|
56 |
**3. The big idea:**
|
57 |
|
58 |
+
Getting pros of both searches made sense and that gave rise to interest in learning sparse representations for queries and documents with some interpretability. The sparse representations also double as implicit or explicit (latent, contextualized) expansion mechanisms for both query and documents. If you are new to query expansion learn more here from the master himself Daniel Tunkelang.
|
59 |
|
60 |
**4. What a Sparse model learns ?**
|
61 |
|
62 |
The model learns to project it's learned dense representations over a MLM head to give a vocabulary distribution. Which is just to say the model can do automatic token expansion. (Image courtesy of pinecone)
|
63 |
|
64 |
+
<img src="./expansion.png" width=600 height=550/>
|
65 |
|
66 |
</details>
|
67 |
|
|
|
76 |
4. Achieves a modest yet competitive effectiveness **MRR@10 37.22** in ID data (& OOD) and a retrieval latency of - **47.27ms**. (multi-threaded) all On **Consumer grade-GPUs** with **only 5 negatives per query**.
|
77 |
4. For Industry setting: Effectiveness on custom domains needs more than just **Trading FLOPS for tiny gains** and The Premise "SPLADE++ are not well suited to mono-cpu retrieval" does not hold.
|
78 |
|
79 |
+
<img src="./ID.png" width=750 height=650/>
|
80 |
|
81 |
*Note: The paper refers to the best performing models as SPLADE++, hence for consistency we are reusing the same.*
|
82 |
|