eval/beir-arguana_ndcg@10 = 0.36665 eval/beir-arguana_recall@100 = 0.96728 eval/beir-avg_ndcg@10 = 0.30767 eval/beir-avg_recall@10 = 0.33533500000000005 eval/beir-avg_recall@100 = 0.523215 eval/beir-avg_recall@20 = 0.383875 eval/beir-climate-fever_ndcg@10 = 0.12105 eval/beir-climate-fever_recall@10 = 0.1548 eval/beir-climate-fever_recall@100 = 0.37817 eval/beir-climate-fever_recall@20 = 0.20165 eval/beir-cqadupstack_ndcg@10 = 0.3031158333333333 eval/beir-cqadupstack_recall@100 = 0.6389233333333334 eval/beir-dbpedia-entity_ndcg@10 = 0.3116 eval/beir-dbpedia-entity_recall@100 = 0.4652 eval/beir-fever_ndcg@10 = 0.63923 eval/beir-fever_recall@100 = 0.93145 eval/beir-fiqa_ndcg@10 = 0.28062 eval/beir-fiqa_recall@100 = 0.61765 eval/beir-hotpotqa_ndcg@10 = 0.49429 eval/beir-hotpotqa_recall@10 = 0.51587 eval/beir-hotpotqa_recall@100 = 0.66826 eval/beir-hotpotqa_recall@20 = 0.5661 eval/beir-msmarco_ndcg@10 = 0.20651 eval/beir-msmarco_recall@100 = 0.67481 eval/beir-nfcorpus_ndcg@10 = 0.34134 eval/beir-nfcorpus_recall@100 = 0.31962 eval/beir-nq_ndcg@10 = 0.27574 eval/beir-nq_recall@100 = 0.7927 eval/beir-quora_ndcg@10 = 0.78549 eval/beir-quora_recall@100 = 0.9752 eval/beir-scidocs_ndcg@10 = 0.15962 eval/beir-scidocs_recall@100 = 0.39283 eval/beir-scifact_ndcg@10 = 0.618 eval/beir-scifact_recall@100 = 0.911 eval/beir-trec-covid_ndcg@10 = 0.58244 eval/beir-trec-covid_recall@100 = 0.10597 eval/beir-webis-touche2020_ndcg@10 = 0.18565 eval/beir-webis-touche2020_recall@100 = 0.43843 eval/qa-curatedtrec-test-acc@100 = 0.9265129682997119 eval/qa-curatedtrec-test-acc@20 = 0.8371757925072046 eval/qa-curatedtrec-test-acc@5 = 0.6354466858789626 eval/qa-entityqs-macro-acc@100 = 0.7567939708496935 eval/qa-entityqs-macro-acc@20 = 0.6403777676502388 eval/qa-entityqs-macro-acc@5 = 0.49428236812619747 eval/qa-nq-test-acc@100 = 0.8024930747922437 eval/qa-nq-test-acc@20 = 0.6501385041551246 eval/qa-nq-test-acc@5 = 0.44155124653739614 eval/qa-squad1-test-acc@100 = 0.7376537369914853 eval/qa-squad1-test-acc@20 = 0.5704824976348155 eval/qa-squad1-test-acc@5 = 0.3709555345316935 eval/qa-trivia-test-acc@100 = 0.8229470520639972 eval/qa-trivia-test-acc@20 = 0.7227083885795103 eval/qa-trivia-test-acc@5 = 0.5689030319101919 eval/qa-webq-test-acc@100 = 0.8061023622047244 eval/qa-webq-test-acc@20 = 0.6520669291338582 eval/qa-webq-test-acc@5 = 0.40994094488188976 eval/senteval-CR = 87.02 eval/senteval-MPQA = 88.67 eval/senteval-MR = 80.75 eval/senteval-MRPC = 70.17 eval/senteval-SICKRelatedness = 0.7122486034563766 eval/senteval-SST2 = 82.8 eval/senteval-STS12 = 0.6520942930121344 eval/senteval-STS13 = 0.7489808796663362 eval/senteval-STS14 = 0.6214541243526541 eval/senteval-STS15 = 0.7497940183329284 eval/senteval-STS16 = 0.7490972681198437 eval/senteval-STSBenchmark = 0.708944371429387 eval/senteval-SUBJ = 95.02 eval/senteval-TREC = 78.63 eval/senteval-avg_sts_7 = 0.7060876511956659 eval/senteval-avg_transfer = 83.2942857142857 train/global_step = 100000