Muennighoff
commited on
Commit
•
88aafd9
1
Parent(s):
ea22fc5
Update README.md (#1)
Browse files- Update README.md (824ac8c53a1599eccd65ca493c1a0565a4d7fb66)
README.md
CHANGED
@@ -15,6 +15,7 @@ model-index:
|
|
15 |
name: MTEB AmazonCounterfactualClassification (en)
|
16 |
config: en
|
17 |
split: test
|
|
|
18 |
metrics:
|
19 |
- type: accuracy
|
20 |
value: 74.07462686567165
|
@@ -29,6 +30,7 @@ model-index:
|
|
29 |
name: MTEB AmazonCounterfactualClassification (de)
|
30 |
config: de
|
31 |
split: test
|
|
|
32 |
metrics:
|
33 |
- type: accuracy
|
34 |
value: 66.63811563169165
|
@@ -43,6 +45,7 @@ model-index:
|
|
43 |
name: MTEB AmazonCounterfactualClassification (en-ext)
|
44 |
config: en-ext
|
45 |
split: test
|
|
|
46 |
metrics:
|
47 |
- type: accuracy
|
48 |
value: 77.21889055472263
|
@@ -57,6 +60,7 @@ model-index:
|
|
57 |
name: MTEB AmazonCounterfactualClassification (ja)
|
58 |
config: ja
|
59 |
split: test
|
|
|
60 |
metrics:
|
61 |
- type: accuracy
|
62 |
value: 58.06209850107067
|
@@ -71,6 +75,7 @@ model-index:
|
|
71 |
name: MTEB AmazonPolarityClassification
|
72 |
config: default
|
73 |
split: test
|
|
|
74 |
metrics:
|
75 |
- type: accuracy
|
76 |
value: 82.30920000000002
|
@@ -85,6 +90,7 @@ model-index:
|
|
85 |
name: MTEB AmazonReviewsClassification (en)
|
86 |
config: en
|
87 |
split: test
|
|
|
88 |
metrics:
|
89 |
- type: accuracy
|
90 |
value: 41.584
|
@@ -97,6 +103,7 @@ model-index:
|
|
97 |
name: MTEB AmazonReviewsClassification (de)
|
98 |
config: de
|
99 |
split: test
|
|
|
100 |
metrics:
|
101 |
- type: accuracy
|
102 |
value: 35.288000000000004
|
@@ -109,6 +116,7 @@ model-index:
|
|
109 |
name: MTEB AmazonReviewsClassification (es)
|
110 |
config: es
|
111 |
split: test
|
|
|
112 |
metrics:
|
113 |
- type: accuracy
|
114 |
value: 38.34
|
@@ -121,6 +129,7 @@ model-index:
|
|
121 |
name: MTEB AmazonReviewsClassification (fr)
|
122 |
config: fr
|
123 |
split: test
|
|
|
124 |
metrics:
|
125 |
- type: accuracy
|
126 |
value: 37.839999999999996
|
@@ -133,6 +142,7 @@ model-index:
|
|
133 |
name: MTEB AmazonReviewsClassification (ja)
|
134 |
config: ja
|
135 |
split: test
|
|
|
136 |
metrics:
|
137 |
- type: accuracy
|
138 |
value: 30.936000000000003
|
@@ -145,6 +155,7 @@ model-index:
|
|
145 |
name: MTEB AmazonReviewsClassification (zh)
|
146 |
config: zh
|
147 |
split: test
|
|
|
148 |
metrics:
|
149 |
- type: accuracy
|
150 |
value: 33.75
|
@@ -157,6 +168,7 @@ model-index:
|
|
157 |
name: MTEB ArguAna
|
158 |
config: default
|
159 |
split: test
|
|
|
160 |
metrics:
|
161 |
- type: map_at_1
|
162 |
value: 13.727
|
@@ -213,6 +225,7 @@ model-index:
|
|
213 |
name: MTEB ArxivClusteringP2P
|
214 |
config: default
|
215 |
split: test
|
|
|
216 |
metrics:
|
217 |
- type: v_measure
|
218 |
value: 40.553923271901695
|
@@ -223,6 +236,7 @@ model-index:
|
|
223 |
name: MTEB ArxivClusteringS2S
|
224 |
config: default
|
225 |
split: test
|
|
|
226 |
metrics:
|
227 |
- type: v_measure
|
228 |
value: 32.49323183712211
|
@@ -233,6 +247,7 @@ model-index:
|
|
233 |
name: MTEB AskUbuntuDupQuestions
|
234 |
config: default
|
235 |
split: test
|
|
|
236 |
metrics:
|
237 |
- type: map
|
238 |
value: 55.89811361443445
|
@@ -245,6 +260,7 @@ model-index:
|
|
245 |
name: MTEB BIOSSES
|
246 |
config: default
|
247 |
split: test
|
|
|
248 |
metrics:
|
249 |
- type: cos_sim_pearson
|
250 |
value: 82.50506557805856
|
@@ -265,6 +281,7 @@ model-index:
|
|
265 |
name: MTEB BUCC (de-en)
|
266 |
config: de-en
|
267 |
split: test
|
|
|
268 |
metrics:
|
269 |
- type: accuracy
|
270 |
value: 75.49060542797494
|
@@ -281,6 +298,7 @@ model-index:
|
|
281 |
name: MTEB BUCC (fr-en)
|
282 |
config: fr-en
|
283 |
split: test
|
|
|
284 |
metrics:
|
285 |
- type: accuracy
|
286 |
value: 0.4182258419546555
|
@@ -297,6 +315,7 @@ model-index:
|
|
297 |
name: MTEB BUCC (ru-en)
|
298 |
config: ru-en
|
299 |
split: test
|
|
|
300 |
metrics:
|
301 |
- type: accuracy
|
302 |
value: 0.013855213023900243
|
@@ -313,6 +332,7 @@ model-index:
|
|
313 |
name: MTEB BUCC (zh-en)
|
314 |
config: zh-en
|
315 |
split: test
|
|
|
316 |
metrics:
|
317 |
- type: accuracy
|
318 |
value: 0.315955766192733
|
@@ -329,6 +349,7 @@ model-index:
|
|
329 |
name: MTEB Banking77Classification
|
330 |
config: default
|
331 |
split: test
|
|
|
332 |
metrics:
|
333 |
- type: accuracy
|
334 |
value: 81.74025974025973
|
@@ -341,6 +362,7 @@ model-index:
|
|
341 |
name: MTEB BiorxivClusteringP2P
|
342 |
config: default
|
343 |
split: test
|
|
|
344 |
metrics:
|
345 |
- type: v_measure
|
346 |
value: 33.59451202614059
|
@@ -351,6 +373,7 @@ model-index:
|
|
351 |
name: MTEB BiorxivClusteringS2S
|
352 |
config: default
|
353 |
split: test
|
|
|
354 |
metrics:
|
355 |
- type: v_measure
|
356 |
value: 29.128241446157165
|
@@ -361,6 +384,7 @@ model-index:
|
|
361 |
name: MTEB CQADupstackAndroidRetrieval
|
362 |
config: default
|
363 |
split: test
|
|
|
364 |
metrics:
|
365 |
- type: map_at_1
|
366 |
value: 26.715
|
@@ -417,6 +441,7 @@ model-index:
|
|
417 |
name: MTEB CQADupstackEnglishRetrieval
|
418 |
config: default
|
419 |
split: test
|
|
|
420 |
metrics:
|
421 |
- type: map_at_1
|
422 |
value: 19.663
|
@@ -473,6 +498,7 @@ model-index:
|
|
473 |
name: MTEB CQADupstackGamingRetrieval
|
474 |
config: default
|
475 |
split: test
|
|
|
476 |
metrics:
|
477 |
- type: map_at_1
|
478 |
value: 30.125
|
@@ -529,6 +555,7 @@ model-index:
|
|
529 |
name: MTEB CQADupstackGisRetrieval
|
530 |
config: default
|
531 |
split: test
|
|
|
532 |
metrics:
|
533 |
- type: map_at_1
|
534 |
value: 16.298000000000002
|
@@ -585,6 +612,7 @@ model-index:
|
|
585 |
name: MTEB CQADupstackMathematicaRetrieval
|
586 |
config: default
|
587 |
split: test
|
|
|
588 |
metrics:
|
589 |
- type: map_at_1
|
590 |
value: 10.958
|
@@ -641,6 +669,7 @@ model-index:
|
|
641 |
name: MTEB CQADupstackPhysicsRetrieval
|
642 |
config: default
|
643 |
split: test
|
|
|
644 |
metrics:
|
645 |
- type: map_at_1
|
646 |
value: 21.217
|
@@ -697,6 +726,7 @@ model-index:
|
|
697 |
name: MTEB CQADupstackProgrammersRetrieval
|
698 |
config: default
|
699 |
split: test
|
|
|
700 |
metrics:
|
701 |
- type: map_at_1
|
702 |
value: 19.274
|
@@ -753,6 +783,7 @@ model-index:
|
|
753 |
name: MTEB CQADupstackRetrieval
|
754 |
config: default
|
755 |
split: test
|
|
|
756 |
metrics:
|
757 |
- type: map_at_1
|
758 |
value: 18.653666666666666
|
@@ -809,6 +840,7 @@ model-index:
|
|
809 |
name: MTEB CQADupstackStatsRetrieval
|
810 |
config: default
|
811 |
split: test
|
|
|
812 |
metrics:
|
813 |
- type: map_at_1
|
814 |
value: 16.07
|
@@ -865,6 +897,7 @@ model-index:
|
|
865 |
name: MTEB CQADupstackTexRetrieval
|
866 |
config: default
|
867 |
split: test
|
|
|
868 |
metrics:
|
869 |
- type: map_at_1
|
870 |
value: 10.847
|
@@ -921,6 +954,7 @@ model-index:
|
|
921 |
name: MTEB CQADupstackUnixRetrieval
|
922 |
config: default
|
923 |
split: test
|
|
|
924 |
metrics:
|
925 |
- type: map_at_1
|
926 |
value: 18.377
|
@@ -977,6 +1011,7 @@ model-index:
|
|
977 |
name: MTEB CQADupstackWebmastersRetrieval
|
978 |
config: default
|
979 |
split: test
|
|
|
980 |
metrics:
|
981 |
- type: map_at_1
|
982 |
value: 20.246
|
@@ -1033,6 +1068,7 @@ model-index:
|
|
1033 |
name: MTEB CQADupstackWordpressRetrieval
|
1034 |
config: default
|
1035 |
split: test
|
|
|
1036 |
metrics:
|
1037 |
- type: map_at_1
|
1038 |
value: 14.054
|
@@ -1089,6 +1125,7 @@ model-index:
|
|
1089 |
name: MTEB ClimateFEVER
|
1090 |
config: default
|
1091 |
split: test
|
|
|
1092 |
metrics:
|
1093 |
- type: map_at_1
|
1094 |
value: 6.122
|
@@ -1145,6 +1182,7 @@ model-index:
|
|
1145 |
name: MTEB DBPedia
|
1146 |
config: default
|
1147 |
split: test
|
|
|
1148 |
metrics:
|
1149 |
- type: map_at_1
|
1150 |
value: 4.672
|
@@ -1201,6 +1239,7 @@ model-index:
|
|
1201 |
name: MTEB EmotionClassification
|
1202 |
config: default
|
1203 |
split: test
|
|
|
1204 |
metrics:
|
1205 |
- type: accuracy
|
1206 |
value: 49.919999999999995
|
@@ -1213,6 +1252,7 @@ model-index:
|
|
1213 |
name: MTEB FEVER
|
1214 |
config: default
|
1215 |
split: test
|
|
|
1216 |
metrics:
|
1217 |
- type: map_at_1
|
1218 |
value: 25.801000000000002
|
@@ -1269,6 +1309,7 @@ model-index:
|
|
1269 |
name: MTEB FiQA2018
|
1270 |
config: default
|
1271 |
split: test
|
|
|
1272 |
metrics:
|
1273 |
- type: map_at_1
|
1274 |
value: 9.142
|
@@ -1325,6 +1366,7 @@ model-index:
|
|
1325 |
name: MTEB HotpotQA
|
1326 |
config: default
|
1327 |
split: test
|
|
|
1328 |
metrics:
|
1329 |
- type: map_at_1
|
1330 |
value: 18.677
|
@@ -1381,6 +1423,7 @@ model-index:
|
|
1381 |
name: MTEB ImdbClassification
|
1382 |
config: default
|
1383 |
split: test
|
|
|
1384 |
metrics:
|
1385 |
- type: accuracy
|
1386 |
value: 74.3292
|
@@ -1395,6 +1438,7 @@ model-index:
|
|
1395 |
name: MTEB MSMARCO
|
1396 |
config: default
|
1397 |
split: validation
|
|
|
1398 |
metrics:
|
1399 |
- type: map_at_1
|
1400 |
value: 6.889000000000001
|
@@ -1451,6 +1495,7 @@ model-index:
|
|
1451 |
name: MTEB MTOPDomainClassification (en)
|
1452 |
config: en
|
1453 |
split: test
|
|
|
1454 |
metrics:
|
1455 |
- type: accuracy
|
1456 |
value: 89.6374829001368
|
@@ -1463,6 +1508,7 @@ model-index:
|
|
1463 |
name: MTEB MTOPDomainClassification (de)
|
1464 |
config: de
|
1465 |
split: test
|
|
|
1466 |
metrics:
|
1467 |
- type: accuracy
|
1468 |
value: 84.54212454212454
|
@@ -1475,6 +1521,7 @@ model-index:
|
|
1475 |
name: MTEB MTOPDomainClassification (es)
|
1476 |
config: es
|
1477 |
split: test
|
|
|
1478 |
metrics:
|
1479 |
- type: accuracy
|
1480 |
value: 86.46430953969313
|
@@ -1487,6 +1534,7 @@ model-index:
|
|
1487 |
name: MTEB MTOPDomainClassification (fr)
|
1488 |
config: fr
|
1489 |
split: test
|
|
|
1490 |
metrics:
|
1491 |
- type: accuracy
|
1492 |
value: 81.31850923896022
|
@@ -1499,6 +1547,7 @@ model-index:
|
|
1499 |
name: MTEB MTOPDomainClassification (hi)
|
1500 |
config: hi
|
1501 |
split: test
|
|
|
1502 |
metrics:
|
1503 |
- type: accuracy
|
1504 |
value: 58.23234134098243
|
@@ -1511,6 +1560,7 @@ model-index:
|
|
1511 |
name: MTEB MTOPDomainClassification (th)
|
1512 |
config: th
|
1513 |
split: test
|
|
|
1514 |
metrics:
|
1515 |
- type: accuracy
|
1516 |
value: 72.28571428571429
|
@@ -1523,6 +1573,7 @@ model-index:
|
|
1523 |
name: MTEB MTOPIntentClassification (en)
|
1524 |
config: en
|
1525 |
split: test
|
|
|
1526 |
metrics:
|
1527 |
- type: accuracy
|
1528 |
value: 70.68171454628363
|
@@ -1535,6 +1586,7 @@ model-index:
|
|
1535 |
name: MTEB MTOPIntentClassification (de)
|
1536 |
config: de
|
1537 |
split: test
|
|
|
1538 |
metrics:
|
1539 |
- type: accuracy
|
1540 |
value: 60.521273598196665
|
@@ -1547,6 +1599,7 @@ model-index:
|
|
1547 |
name: MTEB MTOPIntentClassification (es)
|
1548 |
config: es
|
1549 |
split: test
|
|
|
1550 |
metrics:
|
1551 |
- type: accuracy
|
1552 |
value: 64.32288192128087
|
@@ -1559,6 +1612,7 @@ model-index:
|
|
1559 |
name: MTEB MTOPIntentClassification (fr)
|
1560 |
config: fr
|
1561 |
split: test
|
|
|
1562 |
metrics:
|
1563 |
- type: accuracy
|
1564 |
value: 58.67209520826808
|
@@ -1571,6 +1625,7 @@ model-index:
|
|
1571 |
name: MTEB MTOPIntentClassification (hi)
|
1572 |
config: hi
|
1573 |
split: test
|
|
|
1574 |
metrics:
|
1575 |
- type: accuracy
|
1576 |
value: 41.95769092864826
|
@@ -1583,6 +1638,7 @@ model-index:
|
|
1583 |
name: MTEB MTOPIntentClassification (th)
|
1584 |
config: th
|
1585 |
split: test
|
|
|
1586 |
metrics:
|
1587 |
- type: accuracy
|
1588 |
value: 55.28390596745027
|
@@ -1595,6 +1651,7 @@ model-index:
|
|
1595 |
name: MTEB MassiveIntentClassification (en)
|
1596 |
config: en
|
1597 |
split: test
|
|
|
1598 |
metrics:
|
1599 |
- type: accuracy
|
1600 |
value: 70.00336247478144
|
@@ -1607,6 +1664,7 @@ model-index:
|
|
1607 |
name: MTEB MassiveScenarioClassification (en)
|
1608 |
config: en
|
1609 |
split: test
|
|
|
1610 |
metrics:
|
1611 |
- type: accuracy
|
1612 |
value: 75.0268997982515
|
@@ -1619,6 +1677,7 @@ model-index:
|
|
1619 |
name: MTEB MedrxivClusteringP2P
|
1620 |
config: default
|
1621 |
split: test
|
|
|
1622 |
metrics:
|
1623 |
- type: v_measure
|
1624 |
value: 30.327566856300813
|
@@ -1629,6 +1688,7 @@ model-index:
|
|
1629 |
name: MTEB MedrxivClusteringS2S
|
1630 |
config: default
|
1631 |
split: test
|
|
|
1632 |
metrics:
|
1633 |
- type: v_measure
|
1634 |
value: 28.01650210863619
|
@@ -1639,6 +1699,7 @@ model-index:
|
|
1639 |
name: MTEB MindSmallReranking
|
1640 |
config: default
|
1641 |
split: test
|
|
|
1642 |
metrics:
|
1643 |
- type: map
|
1644 |
value: 31.11041256752524
|
@@ -1651,6 +1712,7 @@ model-index:
|
|
1651 |
name: MTEB NFCorpus
|
1652 |
config: default
|
1653 |
split: test
|
|
|
1654 |
metrics:
|
1655 |
- type: map_at_1
|
1656 |
value: 3.527
|
@@ -1707,6 +1769,7 @@ model-index:
|
|
1707 |
name: MTEB NQ
|
1708 |
config: default
|
1709 |
split: test
|
|
|
1710 |
metrics:
|
1711 |
- type: map_at_1
|
1712 |
value: 11.631
|
@@ -1763,6 +1826,7 @@ model-index:
|
|
1763 |
name: MTEB QuoraRetrieval
|
1764 |
config: default
|
1765 |
split: test
|
|
|
1766 |
metrics:
|
1767 |
- type: map_at_1
|
1768 |
value: 66.64
|
@@ -1819,6 +1883,7 @@ model-index:
|
|
1819 |
name: MTEB RedditClustering
|
1820 |
config: default
|
1821 |
split: test
|
|
|
1822 |
metrics:
|
1823 |
- type: v_measure
|
1824 |
value: 42.17131361041068
|
@@ -1829,6 +1894,7 @@ model-index:
|
|
1829 |
name: MTEB RedditClusteringP2P
|
1830 |
config: default
|
1831 |
split: test
|
|
|
1832 |
metrics:
|
1833 |
- type: v_measure
|
1834 |
value: 48.01815621479994
|
@@ -1839,6 +1905,7 @@ model-index:
|
|
1839 |
name: MTEB SCIDOCS
|
1840 |
config: default
|
1841 |
split: test
|
|
|
1842 |
metrics:
|
1843 |
- type: map_at_1
|
1844 |
value: 3.198
|
@@ -1895,6 +1962,7 @@ model-index:
|
|
1895 |
name: MTEB SICK-R
|
1896 |
config: default
|
1897 |
split: test
|
|
|
1898 |
metrics:
|
1899 |
- type: cos_sim_pearson
|
1900 |
value: 84.5217161312271
|
@@ -1915,6 +1983,7 @@ model-index:
|
|
1915 |
name: MTEB STS12
|
1916 |
config: default
|
1917 |
split: test
|
|
|
1918 |
metrics:
|
1919 |
- type: cos_sim_pearson
|
1920 |
value: 83.70403706922605
|
@@ -1935,6 +2004,7 @@ model-index:
|
|
1935 |
name: MTEB STS13
|
1936 |
config: default
|
1937 |
split: test
|
|
|
1938 |
metrics:
|
1939 |
- type: cos_sim_pearson
|
1940 |
value: 84.94127878986795
|
@@ -1955,6 +2025,7 @@ model-index:
|
|
1955 |
name: MTEB STS14
|
1956 |
config: default
|
1957 |
split: test
|
|
|
1958 |
metrics:
|
1959 |
- type: cos_sim_pearson
|
1960 |
value: 83.1474704168523
|
@@ -1975,6 +2046,7 @@ model-index:
|
|
1975 |
name: MTEB STS15
|
1976 |
config: default
|
1977 |
split: test
|
|
|
1978 |
metrics:
|
1979 |
- type: cos_sim_pearson
|
1980 |
value: 84.81255514055894
|
@@ -1995,6 +2067,7 @@ model-index:
|
|
1995 |
name: MTEB STS16
|
1996 |
config: default
|
1997 |
split: test
|
|
|
1998 |
metrics:
|
1999 |
- type: cos_sim_pearson
|
2000 |
value: 80.63248465157822
|
@@ -2015,6 +2088,7 @@ model-index:
|
|
2015 |
name: MTEB STS17 (en-en)
|
2016 |
config: en-en
|
2017 |
split: test
|
|
|
2018 |
metrics:
|
2019 |
- type: cos_sim_pearson
|
2020 |
value: 90.09066290639687
|
@@ -2035,6 +2109,7 @@ model-index:
|
|
2035 |
name: MTEB STS22 (en)
|
2036 |
config: en
|
2037 |
split: test
|
|
|
2038 |
metrics:
|
2039 |
- type: cos_sim_pearson
|
2040 |
value: 63.52752323046846
|
@@ -2055,6 +2130,7 @@ model-index:
|
|
2055 |
name: MTEB STSBenchmark
|
2056 |
config: default
|
2057 |
split: test
|
|
|
2058 |
metrics:
|
2059 |
- type: cos_sim_pearson
|
2060 |
value: 85.45100366635687
|
@@ -2075,6 +2151,7 @@ model-index:
|
|
2075 |
name: MTEB SciDocsRR
|
2076 |
config: default
|
2077 |
split: test
|
|
|
2078 |
metrics:
|
2079 |
- type: map
|
2080 |
value: 77.53549990038017
|
@@ -2087,6 +2164,7 @@ model-index:
|
|
2087 |
name: MTEB SciFact
|
2088 |
config: default
|
2089 |
split: test
|
|
|
2090 |
metrics:
|
2091 |
- type: map_at_1
|
2092 |
value: 31.167
|
@@ -2143,6 +2221,7 @@ model-index:
|
|
2143 |
name: MTEB SprintDuplicateQuestions
|
2144 |
config: default
|
2145 |
split: test
|
|
|
2146 |
metrics:
|
2147 |
- type: cos_sim_accuracy
|
2148 |
value: 99.55148514851486
|
@@ -2197,6 +2276,7 @@ model-index:
|
|
2197 |
name: MTEB StackExchangeClustering
|
2198 |
config: default
|
2199 |
split: test
|
|
|
2200 |
metrics:
|
2201 |
- type: v_measure
|
2202 |
value: 54.13314692311623
|
@@ -2207,6 +2287,7 @@ model-index:
|
|
2207 |
name: MTEB StackExchangeClusteringP2P
|
2208 |
config: default
|
2209 |
split: test
|
|
|
2210 |
metrics:
|
2211 |
- type: v_measure
|
2212 |
value: 31.115181648287145
|
@@ -2217,6 +2298,7 @@ model-index:
|
|
2217 |
name: MTEB StackOverflowDupQuestions
|
2218 |
config: default
|
2219 |
split: test
|
|
|
2220 |
metrics:
|
2221 |
- type: map
|
2222 |
value: 44.771112666694336
|
@@ -2229,6 +2311,7 @@ model-index:
|
|
2229 |
name: MTEB SummEval
|
2230 |
config: default
|
2231 |
split: test
|
|
|
2232 |
metrics:
|
2233 |
- type: cos_sim_pearson
|
2234 |
value: 30.849429597669374
|
@@ -2245,6 +2328,7 @@ model-index:
|
|
2245 |
name: MTEB TRECCOVID
|
2246 |
config: default
|
2247 |
split: test
|
|
|
2248 |
metrics:
|
2249 |
- type: map_at_1
|
2250 |
value: 0.19499999999999998
|
@@ -2301,6 +2385,7 @@ model-index:
|
|
2301 |
name: MTEB Touche2020
|
2302 |
config: default
|
2303 |
split: test
|
|
|
2304 |
metrics:
|
2305 |
- type: map_at_1
|
2306 |
value: 1.335
|
@@ -2357,6 +2442,7 @@ model-index:
|
|
2357 |
name: MTEB ToxicConversationsClassification
|
2358 |
config: default
|
2359 |
split: test
|
|
|
2360 |
metrics:
|
2361 |
- type: accuracy
|
2362 |
value: 69.93339999999999
|
@@ -2371,6 +2457,7 @@ model-index:
|
|
2371 |
name: MTEB TweetSentimentExtractionClassification
|
2372 |
config: default
|
2373 |
split: test
|
|
|
2374 |
metrics:
|
2375 |
- type: accuracy
|
2376 |
value: 62.43916242218449
|
@@ -2383,6 +2470,7 @@ model-index:
|
|
2383 |
name: MTEB TwentyNewsgroupsClustering
|
2384 |
config: default
|
2385 |
split: test
|
|
|
2386 |
metrics:
|
2387 |
- type: v_measure
|
2388 |
value: 37.202082549859796
|
@@ -2393,6 +2481,7 @@ model-index:
|
|
2393 |
name: MTEB TwitterSemEval2015
|
2394 |
config: default
|
2395 |
split: test
|
|
|
2396 |
metrics:
|
2397 |
- type: cos_sim_accuracy
|
2398 |
value: 83.65023544137807
|
@@ -2447,6 +2536,7 @@ model-index:
|
|
2447 |
name: MTEB TwitterURLCorpus
|
2448 |
config: default
|
2449 |
split: test
|
|
|
2450 |
metrics:
|
2451 |
- type: cos_sim_accuracy
|
2452 |
value: 88.34943920518494
|
|
|
15 |
name: MTEB AmazonCounterfactualClassification (en)
|
16 |
config: en
|
17 |
split: test
|
18 |
+
revision: 2d8a100785abf0ae21420d2a55b0c56e3e1ea996
|
19 |
metrics:
|
20 |
- type: accuracy
|
21 |
value: 74.07462686567165
|
|
|
30 |
name: MTEB AmazonCounterfactualClassification (de)
|
31 |
config: de
|
32 |
split: test
|
33 |
+
revision: 2d8a100785abf0ae21420d2a55b0c56e3e1ea996
|
34 |
metrics:
|
35 |
- type: accuracy
|
36 |
value: 66.63811563169165
|
|
|
45 |
name: MTEB AmazonCounterfactualClassification (en-ext)
|
46 |
config: en-ext
|
47 |
split: test
|
48 |
+
revision: 2d8a100785abf0ae21420d2a55b0c56e3e1ea996
|
49 |
metrics:
|
50 |
- type: accuracy
|
51 |
value: 77.21889055472263
|
|
|
60 |
name: MTEB AmazonCounterfactualClassification (ja)
|
61 |
config: ja
|
62 |
split: test
|
63 |
+
revision: 2d8a100785abf0ae21420d2a55b0c56e3e1ea996
|
64 |
metrics:
|
65 |
- type: accuracy
|
66 |
value: 58.06209850107067
|
|
|
75 |
name: MTEB AmazonPolarityClassification
|
76 |
config: default
|
77 |
split: test
|
78 |
+
revision: 80714f8dcf8cefc218ef4f8c5a966dd83f75a0e1
|
79 |
metrics:
|
80 |
- type: accuracy
|
81 |
value: 82.30920000000002
|
|
|
90 |
name: MTEB AmazonReviewsClassification (en)
|
91 |
config: en
|
92 |
split: test
|
93 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
94 |
metrics:
|
95 |
- type: accuracy
|
96 |
value: 41.584
|
|
|
103 |
name: MTEB AmazonReviewsClassification (de)
|
104 |
config: de
|
105 |
split: test
|
106 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
107 |
metrics:
|
108 |
- type: accuracy
|
109 |
value: 35.288000000000004
|
|
|
116 |
name: MTEB AmazonReviewsClassification (es)
|
117 |
config: es
|
118 |
split: test
|
119 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
120 |
metrics:
|
121 |
- type: accuracy
|
122 |
value: 38.34
|
|
|
129 |
name: MTEB AmazonReviewsClassification (fr)
|
130 |
config: fr
|
131 |
split: test
|
132 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
133 |
metrics:
|
134 |
- type: accuracy
|
135 |
value: 37.839999999999996
|
|
|
142 |
name: MTEB AmazonReviewsClassification (ja)
|
143 |
config: ja
|
144 |
split: test
|
145 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
146 |
metrics:
|
147 |
- type: accuracy
|
148 |
value: 30.936000000000003
|
|
|
155 |
name: MTEB AmazonReviewsClassification (zh)
|
156 |
config: zh
|
157 |
split: test
|
158 |
+
revision: c379a6705fec24a2493fa68e011692605f44e119
|
159 |
metrics:
|
160 |
- type: accuracy
|
161 |
value: 33.75
|
|
|
168 |
name: MTEB ArguAna
|
169 |
config: default
|
170 |
split: test
|
171 |
+
revision: 5b3e3697907184a9b77a3c99ee9ea1a9cbb1e4e3
|
172 |
metrics:
|
173 |
- type: map_at_1
|
174 |
value: 13.727
|
|
|
225 |
name: MTEB ArxivClusteringP2P
|
226 |
config: default
|
227 |
split: test
|
228 |
+
revision: 0bbdb47bcbe3a90093699aefeed338a0f28a7ee8
|
229 |
metrics:
|
230 |
- type: v_measure
|
231 |
value: 40.553923271901695
|
|
|
236 |
name: MTEB ArxivClusteringS2S
|
237 |
config: default
|
238 |
split: test
|
239 |
+
revision: b73bd54100e5abfa6e3a23dcafb46fe4d2438dc3
|
240 |
metrics:
|
241 |
- type: v_measure
|
242 |
value: 32.49323183712211
|
|
|
247 |
name: MTEB AskUbuntuDupQuestions
|
248 |
config: default
|
249 |
split: test
|
250 |
+
revision: 4d853f94cd57d85ec13805aeeac3ae3e5eb4c49c
|
251 |
metrics:
|
252 |
- type: map
|
253 |
value: 55.89811361443445
|
|
|
260 |
name: MTEB BIOSSES
|
261 |
config: default
|
262 |
split: test
|
263 |
+
revision: 9ee918f184421b6bd48b78f6c714d86546106103
|
264 |
metrics:
|
265 |
- type: cos_sim_pearson
|
266 |
value: 82.50506557805856
|
|
|
281 |
name: MTEB BUCC (de-en)
|
282 |
config: de-en
|
283 |
split: test
|
284 |
+
revision: d51519689f32196a32af33b075a01d0e7c51e252
|
285 |
metrics:
|
286 |
- type: accuracy
|
287 |
value: 75.49060542797494
|
|
|
298 |
name: MTEB BUCC (fr-en)
|
299 |
config: fr-en
|
300 |
split: test
|
301 |
+
revision: d51519689f32196a32af33b075a01d0e7c51e252
|
302 |
metrics:
|
303 |
- type: accuracy
|
304 |
value: 0.4182258419546555
|
|
|
315 |
name: MTEB BUCC (ru-en)
|
316 |
config: ru-en
|
317 |
split: test
|
318 |
+
revision: d51519689f32196a32af33b075a01d0e7c51e252
|
319 |
metrics:
|
320 |
- type: accuracy
|
321 |
value: 0.013855213023900243
|
|
|
332 |
name: MTEB BUCC (zh-en)
|
333 |
config: zh-en
|
334 |
split: test
|
335 |
+
revision: d51519689f32196a32af33b075a01d0e7c51e252
|
336 |
metrics:
|
337 |
- type: accuracy
|
338 |
value: 0.315955766192733
|
|
|
349 |
name: MTEB Banking77Classification
|
350 |
config: default
|
351 |
split: test
|
352 |
+
revision: 44fa15921b4c889113cc5df03dd4901b49161ab7
|
353 |
metrics:
|
354 |
- type: accuracy
|
355 |
value: 81.74025974025973
|
|
|
362 |
name: MTEB BiorxivClusteringP2P
|
363 |
config: default
|
364 |
split: test
|
365 |
+
revision: 11d0121201d1f1f280e8cc8f3d98fb9c4d9f9c55
|
366 |
metrics:
|
367 |
- type: v_measure
|
368 |
value: 33.59451202614059
|
|
|
373 |
name: MTEB BiorxivClusteringS2S
|
374 |
config: default
|
375 |
split: test
|
376 |
+
revision: c0fab014e1bcb8d3a5e31b2088972a1e01547dc1
|
377 |
metrics:
|
378 |
- type: v_measure
|
379 |
value: 29.128241446157165
|
|
|
384 |
name: MTEB CQADupstackAndroidRetrieval
|
385 |
config: default
|
386 |
split: test
|
387 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
388 |
metrics:
|
389 |
- type: map_at_1
|
390 |
value: 26.715
|
|
|
441 |
name: MTEB CQADupstackEnglishRetrieval
|
442 |
config: default
|
443 |
split: test
|
444 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
445 |
metrics:
|
446 |
- type: map_at_1
|
447 |
value: 19.663
|
|
|
498 |
name: MTEB CQADupstackGamingRetrieval
|
499 |
config: default
|
500 |
split: test
|
501 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
502 |
metrics:
|
503 |
- type: map_at_1
|
504 |
value: 30.125
|
|
|
555 |
name: MTEB CQADupstackGisRetrieval
|
556 |
config: default
|
557 |
split: test
|
558 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
559 |
metrics:
|
560 |
- type: map_at_1
|
561 |
value: 16.298000000000002
|
|
|
612 |
name: MTEB CQADupstackMathematicaRetrieval
|
613 |
config: default
|
614 |
split: test
|
615 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
616 |
metrics:
|
617 |
- type: map_at_1
|
618 |
value: 10.958
|
|
|
669 |
name: MTEB CQADupstackPhysicsRetrieval
|
670 |
config: default
|
671 |
split: test
|
672 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
673 |
metrics:
|
674 |
- type: map_at_1
|
675 |
value: 21.217
|
|
|
726 |
name: MTEB CQADupstackProgrammersRetrieval
|
727 |
config: default
|
728 |
split: test
|
729 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
730 |
metrics:
|
731 |
- type: map_at_1
|
732 |
value: 19.274
|
|
|
783 |
name: MTEB CQADupstackRetrieval
|
784 |
config: default
|
785 |
split: test
|
786 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
787 |
metrics:
|
788 |
- type: map_at_1
|
789 |
value: 18.653666666666666
|
|
|
840 |
name: MTEB CQADupstackStatsRetrieval
|
841 |
config: default
|
842 |
split: test
|
843 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
844 |
metrics:
|
845 |
- type: map_at_1
|
846 |
value: 16.07
|
|
|
897 |
name: MTEB CQADupstackTexRetrieval
|
898 |
config: default
|
899 |
split: test
|
900 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
901 |
metrics:
|
902 |
- type: map_at_1
|
903 |
value: 10.847
|
|
|
954 |
name: MTEB CQADupstackUnixRetrieval
|
955 |
config: default
|
956 |
split: test
|
957 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
958 |
metrics:
|
959 |
- type: map_at_1
|
960 |
value: 18.377
|
|
|
1011 |
name: MTEB CQADupstackWebmastersRetrieval
|
1012 |
config: default
|
1013 |
split: test
|
1014 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
1015 |
metrics:
|
1016 |
- type: map_at_1
|
1017 |
value: 20.246
|
|
|
1068 |
name: MTEB CQADupstackWordpressRetrieval
|
1069 |
config: default
|
1070 |
split: test
|
1071 |
+
revision: 2b9f5791698b5be7bc5e10535c8690f20043c3db
|
1072 |
metrics:
|
1073 |
- type: map_at_1
|
1074 |
value: 14.054
|
|
|
1125 |
name: MTEB ClimateFEVER
|
1126 |
config: default
|
1127 |
split: test
|
1128 |
+
revision: 392b78eb68c07badcd7c2cd8f39af108375dfcce
|
1129 |
metrics:
|
1130 |
- type: map_at_1
|
1131 |
value: 6.122
|
|
|
1182 |
name: MTEB DBPedia
|
1183 |
config: default
|
1184 |
split: test
|
1185 |
+
revision: f097057d03ed98220bc7309ddb10b71a54d667d6
|
1186 |
metrics:
|
1187 |
- type: map_at_1
|
1188 |
value: 4.672
|
|
|
1239 |
name: MTEB EmotionClassification
|
1240 |
config: default
|
1241 |
split: test
|
1242 |
+
revision: 829147f8f75a25f005913200eb5ed41fae320aa1
|
1243 |
metrics:
|
1244 |
- type: accuracy
|
1245 |
value: 49.919999999999995
|
|
|
1252 |
name: MTEB FEVER
|
1253 |
config: default
|
1254 |
split: test
|
1255 |
+
revision: 1429cf27e393599b8b359b9b72c666f96b2525f9
|
1256 |
metrics:
|
1257 |
- type: map_at_1
|
1258 |
value: 25.801000000000002
|
|
|
1309 |
name: MTEB FiQA2018
|
1310 |
config: default
|
1311 |
split: test
|
1312 |
+
revision: 41b686a7f28c59bcaaa5791efd47c67c8ebe28be
|
1313 |
metrics:
|
1314 |
- type: map_at_1
|
1315 |
value: 9.142
|
|
|
1366 |
name: MTEB HotpotQA
|
1367 |
config: default
|
1368 |
split: test
|
1369 |
+
revision: 766870b35a1b9ca65e67a0d1913899973551fc6c
|
1370 |
metrics:
|
1371 |
- type: map_at_1
|
1372 |
value: 18.677
|
|
|
1423 |
name: MTEB ImdbClassification
|
1424 |
config: default
|
1425 |
split: test
|
1426 |
+
revision: 8d743909f834c38949e8323a8a6ce8721ea6c7f4
|
1427 |
metrics:
|
1428 |
- type: accuracy
|
1429 |
value: 74.3292
|
|
|
1438 |
name: MTEB MSMARCO
|
1439 |
config: default
|
1440 |
split: validation
|
1441 |
+
revision: e6838a846e2408f22cf5cc337ebc83e0bcf77849
|
1442 |
metrics:
|
1443 |
- type: map_at_1
|
1444 |
value: 6.889000000000001
|
|
|
1495 |
name: MTEB MTOPDomainClassification (en)
|
1496 |
config: en
|
1497 |
split: test
|
1498 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1499 |
metrics:
|
1500 |
- type: accuracy
|
1501 |
value: 89.6374829001368
|
|
|
1508 |
name: MTEB MTOPDomainClassification (de)
|
1509 |
config: de
|
1510 |
split: test
|
1511 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1512 |
metrics:
|
1513 |
- type: accuracy
|
1514 |
value: 84.54212454212454
|
|
|
1521 |
name: MTEB MTOPDomainClassification (es)
|
1522 |
config: es
|
1523 |
split: test
|
1524 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1525 |
metrics:
|
1526 |
- type: accuracy
|
1527 |
value: 86.46430953969313
|
|
|
1534 |
name: MTEB MTOPDomainClassification (fr)
|
1535 |
config: fr
|
1536 |
split: test
|
1537 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1538 |
metrics:
|
1539 |
- type: accuracy
|
1540 |
value: 81.31850923896022
|
|
|
1547 |
name: MTEB MTOPDomainClassification (hi)
|
1548 |
config: hi
|
1549 |
split: test
|
1550 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1551 |
metrics:
|
1552 |
- type: accuracy
|
1553 |
value: 58.23234134098243
|
|
|
1560 |
name: MTEB MTOPDomainClassification (th)
|
1561 |
config: th
|
1562 |
split: test
|
1563 |
+
revision: a7e2a951126a26fc8c6a69f835f33a346ba259e3
|
1564 |
metrics:
|
1565 |
- type: accuracy
|
1566 |
value: 72.28571428571429
|
|
|
1573 |
name: MTEB MTOPIntentClassification (en)
|
1574 |
config: en
|
1575 |
split: test
|
1576 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1577 |
metrics:
|
1578 |
- type: accuracy
|
1579 |
value: 70.68171454628363
|
|
|
1586 |
name: MTEB MTOPIntentClassification (de)
|
1587 |
config: de
|
1588 |
split: test
|
1589 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1590 |
metrics:
|
1591 |
- type: accuracy
|
1592 |
value: 60.521273598196665
|
|
|
1599 |
name: MTEB MTOPIntentClassification (es)
|
1600 |
config: es
|
1601 |
split: test
|
1602 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1603 |
metrics:
|
1604 |
- type: accuracy
|
1605 |
value: 64.32288192128087
|
|
|
1612 |
name: MTEB MTOPIntentClassification (fr)
|
1613 |
config: fr
|
1614 |
split: test
|
1615 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1616 |
metrics:
|
1617 |
- type: accuracy
|
1618 |
value: 58.67209520826808
|
|
|
1625 |
name: MTEB MTOPIntentClassification (hi)
|
1626 |
config: hi
|
1627 |
split: test
|
1628 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1629 |
metrics:
|
1630 |
- type: accuracy
|
1631 |
value: 41.95769092864826
|
|
|
1638 |
name: MTEB MTOPIntentClassification (th)
|
1639 |
config: th
|
1640 |
split: test
|
1641 |
+
revision: 6299947a7777084cc2d4b64235bf7190381ce755
|
1642 |
metrics:
|
1643 |
- type: accuracy
|
1644 |
value: 55.28390596745027
|
|
|
1651 |
name: MTEB MassiveIntentClassification (en)
|
1652 |
config: en
|
1653 |
split: test
|
1654 |
+
revision: 072a486a144adf7f4479a4a0dddb2152e161e1ea
|
1655 |
metrics:
|
1656 |
- type: accuracy
|
1657 |
value: 70.00336247478144
|
|
|
1664 |
name: MTEB MassiveScenarioClassification (en)
|
1665 |
config: en
|
1666 |
split: test
|
1667 |
+
revision: 7d571f92784cd94a019292a1f45445077d0ef634
|
1668 |
metrics:
|
1669 |
- type: accuracy
|
1670 |
value: 75.0268997982515
|
|
|
1677 |
name: MTEB MedrxivClusteringP2P
|
1678 |
config: default
|
1679 |
split: test
|
1680 |
+
revision: dcefc037ef84348e49b0d29109e891c01067226b
|
1681 |
metrics:
|
1682 |
- type: v_measure
|
1683 |
value: 30.327566856300813
|
|
|
1688 |
name: MTEB MedrxivClusteringS2S
|
1689 |
config: default
|
1690 |
split: test
|
1691 |
+
revision: 3cd0e71dfbe09d4de0f9e5ecba43e7ce280959dc
|
1692 |
metrics:
|
1693 |
- type: v_measure
|
1694 |
value: 28.01650210863619
|
|
|
1699 |
name: MTEB MindSmallReranking
|
1700 |
config: default
|
1701 |
split: test
|
1702 |
+
revision: 3bdac13927fdc888b903db93b2ffdbd90b295a69
|
1703 |
metrics:
|
1704 |
- type: map
|
1705 |
value: 31.11041256752524
|
|
|
1712 |
name: MTEB NFCorpus
|
1713 |
config: default
|
1714 |
split: test
|
1715 |
+
revision: 7eb63cc0c1eb59324d709ebed25fcab851fa7610
|
1716 |
metrics:
|
1717 |
- type: map_at_1
|
1718 |
value: 3.527
|
|
|
1769 |
name: MTEB NQ
|
1770 |
config: default
|
1771 |
split: test
|
1772 |
+
revision: 6062aefc120bfe8ece5897809fb2e53bfe0d128c
|
1773 |
metrics:
|
1774 |
- type: map_at_1
|
1775 |
value: 11.631
|
|
|
1826 |
name: MTEB QuoraRetrieval
|
1827 |
config: default
|
1828 |
split: test
|
1829 |
+
revision: 6205996560df11e3a3da9ab4f926788fc30a7db4
|
1830 |
metrics:
|
1831 |
- type: map_at_1
|
1832 |
value: 66.64
|
|
|
1883 |
name: MTEB RedditClustering
|
1884 |
config: default
|
1885 |
split: test
|
1886 |
+
revision: b2805658ae38990172679479369a78b86de8c390
|
1887 |
metrics:
|
1888 |
- type: v_measure
|
1889 |
value: 42.17131361041068
|
|
|
1894 |
name: MTEB RedditClusteringP2P
|
1895 |
config: default
|
1896 |
split: test
|
1897 |
+
revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
|
1898 |
metrics:
|
1899 |
- type: v_measure
|
1900 |
value: 48.01815621479994
|
|
|
1905 |
name: MTEB SCIDOCS
|
1906 |
config: default
|
1907 |
split: test
|
1908 |
+
revision: 5c59ef3e437a0a9651c8fe6fde943e7dce59fba5
|
1909 |
metrics:
|
1910 |
- type: map_at_1
|
1911 |
value: 3.198
|
|
|
1962 |
name: MTEB SICK-R
|
1963 |
config: default
|
1964 |
split: test
|
1965 |
+
revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
|
1966 |
metrics:
|
1967 |
- type: cos_sim_pearson
|
1968 |
value: 84.5217161312271
|
|
|
1983 |
name: MTEB STS12
|
1984 |
config: default
|
1985 |
split: test
|
1986 |
+
revision: fdf84275bb8ce4b49c971d02e84dd1abc677a50f
|
1987 |
metrics:
|
1988 |
- type: cos_sim_pearson
|
1989 |
value: 83.70403706922605
|
|
|
2004 |
name: MTEB STS13
|
2005 |
config: default
|
2006 |
split: test
|
2007 |
+
revision: 1591bfcbe8c69d4bf7fe2a16e2451017832cafb9
|
2008 |
metrics:
|
2009 |
- type: cos_sim_pearson
|
2010 |
value: 84.94127878986795
|
|
|
2025 |
name: MTEB STS14
|
2026 |
config: default
|
2027 |
split: test
|
2028 |
+
revision: e2125984e7df8b7871f6ae9949cf6b6795e7c54b
|
2029 |
metrics:
|
2030 |
- type: cos_sim_pearson
|
2031 |
value: 83.1474704168523
|
|
|
2046 |
name: MTEB STS15
|
2047 |
config: default
|
2048 |
split: test
|
2049 |
+
revision: 1cd7298cac12a96a373b6a2f18738bb3e739a9b6
|
2050 |
metrics:
|
2051 |
- type: cos_sim_pearson
|
2052 |
value: 84.81255514055894
|
|
|
2067 |
name: MTEB STS16
|
2068 |
config: default
|
2069 |
split: test
|
2070 |
+
revision: 360a0b2dff98700d09e634a01e1cc1624d3e42cd
|
2071 |
metrics:
|
2072 |
- type: cos_sim_pearson
|
2073 |
value: 80.63248465157822
|
|
|
2088 |
name: MTEB STS17 (en-en)
|
2089 |
config: en-en
|
2090 |
split: test
|
2091 |
+
revision: 9fc37e8c632af1c87a3d23e685d49552a02582a0
|
2092 |
metrics:
|
2093 |
- type: cos_sim_pearson
|
2094 |
value: 90.09066290639687
|
|
|
2109 |
name: MTEB STS22 (en)
|
2110 |
config: en
|
2111 |
split: test
|
2112 |
+
revision: 2de6ce8c1921b71a755b262c6b57fef195dd7906
|
2113 |
metrics:
|
2114 |
- type: cos_sim_pearson
|
2115 |
value: 63.52752323046846
|
|
|
2130 |
name: MTEB STSBenchmark
|
2131 |
config: default
|
2132 |
split: test
|
2133 |
+
revision: 8913289635987208e6e7c72789e4be2fe94b6abd
|
2134 |
metrics:
|
2135 |
- type: cos_sim_pearson
|
2136 |
value: 85.45100366635687
|
|
|
2151 |
name: MTEB SciDocsRR
|
2152 |
config: default
|
2153 |
split: test
|
2154 |
+
revision: 56a6d0140cf6356659e2a7c1413286a774468d44
|
2155 |
metrics:
|
2156 |
- type: map
|
2157 |
value: 77.53549990038017
|
|
|
2164 |
name: MTEB SciFact
|
2165 |
config: default
|
2166 |
split: test
|
2167 |
+
revision: a75ae049398addde9b70f6b268875f5cbce99089
|
2168 |
metrics:
|
2169 |
- type: map_at_1
|
2170 |
value: 31.167
|
|
|
2221 |
name: MTEB SprintDuplicateQuestions
|
2222 |
config: default
|
2223 |
split: test
|
2224 |
+
revision: 5a8256d0dff9c4bd3be3ba3e67e4e70173f802ea
|
2225 |
metrics:
|
2226 |
- type: cos_sim_accuracy
|
2227 |
value: 99.55148514851486
|
|
|
2276 |
name: MTEB StackExchangeClustering
|
2277 |
config: default
|
2278 |
split: test
|
2279 |
+
revision: 70a89468f6dccacc6aa2b12a6eac54e74328f235
|
2280 |
metrics:
|
2281 |
- type: v_measure
|
2282 |
value: 54.13314692311623
|
|
|
2287 |
name: MTEB StackExchangeClusteringP2P
|
2288 |
config: default
|
2289 |
split: test
|
2290 |
+
revision: d88009ab563dd0b16cfaf4436abaf97fa3550cf0
|
2291 |
metrics:
|
2292 |
- type: v_measure
|
2293 |
value: 31.115181648287145
|
|
|
2298 |
name: MTEB StackOverflowDupQuestions
|
2299 |
config: default
|
2300 |
split: test
|
2301 |
+
revision: ef807ea29a75ec4f91b50fd4191cb4ee4589a9f9
|
2302 |
metrics:
|
2303 |
- type: map
|
2304 |
value: 44.771112666694336
|
|
|
2311 |
name: MTEB SummEval
|
2312 |
config: default
|
2313 |
split: test
|
2314 |
+
revision: 8753c2788d36c01fc6f05d03fe3f7268d63f9122
|
2315 |
metrics:
|
2316 |
- type: cos_sim_pearson
|
2317 |
value: 30.849429597669374
|
|
|
2328 |
name: MTEB TRECCOVID
|
2329 |
config: default
|
2330 |
split: test
|
2331 |
+
revision: 2c8041b2c07a79b6f7ba8fe6acc72e5d9f92d217
|
2332 |
metrics:
|
2333 |
- type: map_at_1
|
2334 |
value: 0.19499999999999998
|
|
|
2385 |
name: MTEB Touche2020
|
2386 |
config: default
|
2387 |
split: test
|
2388 |
+
revision: 527b7d77e16e343303e68cb6af11d6e18b9f7b3b
|
2389 |
metrics:
|
2390 |
- type: map_at_1
|
2391 |
value: 1.335
|
|
|
2442 |
name: MTEB ToxicConversationsClassification
|
2443 |
config: default
|
2444 |
split: test
|
2445 |
+
revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
|
2446 |
metrics:
|
2447 |
- type: accuracy
|
2448 |
value: 69.93339999999999
|
|
|
2457 |
name: MTEB TweetSentimentExtractionClassification
|
2458 |
config: default
|
2459 |
split: test
|
2460 |
+
revision: 62146448f05be9e52a36b8ee9936447ea787eede
|
2461 |
metrics:
|
2462 |
- type: accuracy
|
2463 |
value: 62.43916242218449
|
|
|
2470 |
name: MTEB TwentyNewsgroupsClustering
|
2471 |
config: default
|
2472 |
split: test
|
2473 |
+
revision: 091a54f9a36281ce7d6590ec8c75dd485e7e01d4
|
2474 |
metrics:
|
2475 |
- type: v_measure
|
2476 |
value: 37.202082549859796
|
|
|
2481 |
name: MTEB TwitterSemEval2015
|
2482 |
config: default
|
2483 |
split: test
|
2484 |
+
revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
|
2485 |
metrics:
|
2486 |
- type: cos_sim_accuracy
|
2487 |
value: 83.65023544137807
|
|
|
2536 |
name: MTEB TwitterURLCorpus
|
2537 |
config: default
|
2538 |
split: test
|
2539 |
+
revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
|
2540 |
metrics:
|
2541 |
- type: cos_sim_accuracy
|
2542 |
value: 88.34943920518494
|