Update README.md
Browse files
README.md
CHANGED
@@ -2614,7 +2614,7 @@ The models are built upon the `transformer++` encoder [backbone](https://hugging
|
|
2614 |
The `gte-v1.5` series achieve state-of-the-art scores on the MTEB benchmark within the same model size category and prodvide competitive on the LoCo long-context retrieval tests (refer to [Evaluation](#evaluation)).
|
2615 |
|
2616 |
We also present the [`gte-Qwen1.5-7B-instruct`](https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct),
|
2617 |
-
a SOTA instruction-tuned
|
2618 |
|
2619 |
<!-- Provide a longer summary of what this model is. -->
|
2620 |
|
@@ -2628,7 +2628,7 @@ a SOTA instruction-tuned bilingual embedding model that ranked 2nd in MTEB and 1
|
|
2628 |
|
2629 |
| Models | Language | Model Size | Max Seq. Length | Dimension | MTEB-en | LoCo |
|
2630 |
|:-----: | :-----: |:-----: |:-----: |:-----: | :-----: | :-----: |
|
2631 |
-
|[`gte-Qwen1.5-7B-instruct`](https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct)|
|
2632 |
|[`gte-large-en-v1.5`](https://huggingface.co/Alibaba-NLP/gte-large-en-v1.5) | English | 434 | 8192 | 1024 | 65.39 | 86.71 |
|
2633 |
|[`gte-base-en-v1.5`](https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5) | English | 137 | 8192 | 768 | 64.11 | 87.44 |
|
2634 |
|
|
|
2614 |
The `gte-v1.5` series achieve state-of-the-art scores on the MTEB benchmark within the same model size category and prodvide competitive on the LoCo long-context retrieval tests (refer to [Evaluation](#evaluation)).
|
2615 |
|
2616 |
We also present the [`gte-Qwen1.5-7B-instruct`](https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct),
|
2617 |
+
a SOTA instruction-tuned multi-lingual embedding model that ranked 2nd in MTEB and 1st in C-MTEB.
|
2618 |
|
2619 |
<!-- Provide a longer summary of what this model is. -->
|
2620 |
|
|
|
2628 |
|
2629 |
| Models | Language | Model Size | Max Seq. Length | Dimension | MTEB-en | LoCo |
|
2630 |
|:-----: | :-----: |:-----: |:-----: |:-----: | :-----: | :-----: |
|
2631 |
+
|[`gte-Qwen1.5-7B-instruct`](https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct)| Multiple | 7720 | 32768 | 4096 | 67.34 | 87.57 |
|
2632 |
|[`gte-large-en-v1.5`](https://huggingface.co/Alibaba-NLP/gte-large-en-v1.5) | English | 434 | 8192 | 1024 | 65.39 | 86.71 |
|
2633 |
|[`gte-base-en-v1.5`](https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5) | English | 137 | 8192 | 768 | 64.11 | 87.44 |
|
2634 |
|