{ "dataset_revision": "59042f120c80e8afa9cdbb224f67076cec0fc9a7", "task_name": "MindSmallReranking", "mteb_version": "1.25.1", "scores": { "test": [ { "map": 0.306437, "mrr": 0.316768, "nAUC_map_max": -0.182556, "nAUC_map_std": -0.016765, "nAUC_map_diff1": 0.134889, "nAUC_mrr_max": -0.128279, "nAUC_mrr_std": 0.003295, "nAUC_mrr_diff1": 0.128104, "main_score": 0.306437, "hf_subset": "default", "languages": [ "eng-Latn" ] } ] }, "evaluation_time": 120.96634793281555, "kg_co2_emissions": null }